- 1Center of Excellence in Intelligent Engineering Systems, King Abdulaziz University, Jeddah, Saudi Arabia
- 2Electrical and Computer Engineering Department, King Abdulaziz University, Jeddah, Saudi Arabia
- 3Department of Computer Engineering, Chosun University, Gwangju, South Korea
- 4Research and Development, Love For Data, Karachi, Pakistan
- 5School of Electrical, Electronic and Computer Engineering, The University of Western Australia, Perth, WA, Australia
- 6College of Engineering, Karachi Institute of Economics and Technology, Korangi Creek, Karachi, Pakistan
- 7Department of Radiology, Faculty of Medicine, King Abdulaziz University, Jeddah, Saudi Arabia
- 8Department of Bio and Brain Engineering, Daejeon, South Korea
Extracelluar matrix (ECM) proteins create complex networks of macromolecules which fill-in the extracellular spaces of living tissues. They provide structural support and play an important role in maintaining cellular functions. Identification of ECM proteins can play a vital role in studying various types of diseases. Conventional wet lab–based methods are reliable; however, they are expensive and time consuming and are, therefore, not scalable. In this research, we propose a sequence-based novel machine learning approach for the prediction of ECM proteins. In the proposed method, composition of k-spaced amino acid pair (CKSAAP) features are encoded into a classifiable latent space (LS) with the help of deep latent space encoding (LSE). A comprehensive ablation analysis is conducted for performance evaluation of the proposed method. Results are compared with other state-of-the-art methods on the benchmark dataset, and the proposed ECM-LSE approach has shown to comprehensively outperform the contemporary methods.
1 Introduction
Extracelluar matrix (ECM) is a network of fibrous proteins filled in the extracellular spaces of living tissues to provide structural support for the cells (Karagöz et al., 2021). It is significant for cell functionality and plays an important role in the physiological dynamics. ECMs are also responsible for the promotion of vital cellular processes, including differentiation, adhesion, proliferation, apoptosis, and migration (Klavert and van der Eerden, 2021; Hiraki et al., 2021; Mathews et al., 2012; Endo et al., 2012; Kim et al., 2011). The chemical composition of ECM mainly consists of minerals, proteoglycans, proteins, and water. The proteins in ECM act more like a fibrous material which gives strength to the cells. Several studies have demonstrated that the mutation in the ECM genes can cause severe adverse effects in the cell structure resulting in a number of diseases, including arthritis and cancer (Kizawa et al., 2005; Hu et al., 2007).
Functional research on ECM protein has resulted in the development of useful biomaterials which are used in many fields of medicine, such as tissue engineering and cell therapy (Ma et al., 2019; Gonzalez-Pujana et al., 2019). Proteins, in general, are active elements and play a variety of roles depending on their residing location in a cell. Likewise, the functionality of the ECM varies with the change in the proteins. The problem of protein localization is therefore considered to be an important step toward the understanding of protein functionality (Horton et al., 2007). Identification of subcellular location is however considered to be a nontrivial task and requires extensive experimentation which is prohibitively expensive. Therefore, a variety of computational methods have been developed to facilitate the process (Ras-Carmona et al., 2021; Wang et al., 2021; Chou, 2011). In particular, for different species of plants, animals, and microorganisms, a number of useful techniques have been explored (Zhao et al., 2021; Hou et al., 2021; Chou et al., 2012; Otzen et al., 2021; Wu et al., 2011; Asim et al., 2021; Xiao et al., 2011; Shen et al., 2021; Wu et al., 2012; Lewis et al., 2014). Bioinformatics methods, with the aid of machine learning algorithms, have demonstrated adequate performance for a variety of applications. A detailed review of computational methods to classify secreted proteins has been provided by Klee and Sosa (2007). Typically, three aspects are focused on the development of a computational method: 1) feature extraction—in which the peptide sequence is translated/encoded into a numerical format to make them readable by the model, 2) feature selection—which is concerned with the removal of the redundant information from the feature space and results in the model’s robustness, and 3) model construction and evaluation—which includes development of a prediction model, followed by training and testing steps to evaluate performance.
The first benchmark in-silico approach to predict the extracellular proteins was presented by Jung et al. (2010)and was named as ECM protein prediction (ECMPP). The research used the feature augmentation method and crafted a feature set of 91 attributes. One of the limitations of the study was the use of a small dataset for performance evaluation; also, only the receiver-operating characteristics (ROC) were used for the performance evaluation. Since then, many researchers have paid attention toward the development of machine learning methods for ECM prediction. As extracellular matrix proteins are linked to the outer surface of the cell, they have close association with its secretory mechanism and are naturally associated with the secretary proteins. Therefore, it is reasonable to consider possible ECM candidates as a subset of secretary proteins (Kandaswamy et al., 2010; Bendtsen et al., 2004; Horton et al., 2006). Based on this knowledge, Kandaswamy et al. (2013) improved the ECM prediction method and presented EcmPred. EcmPred (Kandaswamy et al., 2013) used a random forest (RF)–based classifier which was trained on the combination of sequence-derived properties of the proteins including individual and group frequencies of amino acids with the physicochemical properties. Another method named prediction of ECM (PECM) (Zhang et al., 2014) utilized a handcrafted feature set designed by the combination of the most discriminative attributes of the protein sequences including evolutionary and structural information as well as the physicochemical properties of the peptide sequences. An incremental feature selection (IFS) method was employed for the selection of optimal features which were used to train a support vector machine (SVM)–based classifier. Several other methods have also been proposed to serve the task of ECM prediction. None of them, however, focuses on the encoding of sequence-driven feature into a classifiable latent-space (LS). The primary objective of latent space–based learning is to design a reduced feature space for clustering of proteins. The LS is, therefore, a representation of the input signal in a reduced space. The latent-space encoding (LSE) is based on an assumption of a low-rank input (i.e. highly redundant) which can be compressed to a low dimensional signal using LSE. The process is considered to be reversible as the original signal could be reconstructed from the LS. The details of LS and LSE have been provided in the Subsection 2.4.
Development of a feature space and selection of the best features are fundamental steps in designing machine learning models (Lyu et al., 2021). In particular, for the protein sequence classification task, a variety of feature extraction techniques have been proposed including amino acid composition (AAC), dipeptide composition (DPC), N-segmented sequence features, physicochemical composition, and secondary structure features (Naseem et al., 2017; Khan et al., 2018; Kandaswamy et al., 2011). The sole purpose of each feature extraction technique is to encode maximum useful information from a variable length protein sequence into a fixed-sized vector. In the recent past, inspired by the success of deep long short-term memory (LSTM) models, some approaches similar to word2vec (Mikolov et al., 2013) have been proposed to successfully learn latent space encoding directly from variable length sequences (Ding et al., 2019). The direct sequence to latent space encoding method produces good generalization models (Zemouri, 2020); however, they usually rely on the availability of a large training dataset. Furthermore, the direct extraction of latent space features from a limited number of sequences such as, bioluminescence (Zhang et al., 2021), antioxidant (Olsen et al., 2020), ECM (Kabir et al., 2018), antifreeze proteins (AFPs) (Kandaswamy et al., 2011), or other classes of proteins is a challenging problem. In this study, we propose a hybrid approach where all proteins are first encoded into a large feature set obtained through composition of k-spaced amino acid pair encoding. A latent space representation of composition of k-spaced amino acid pairs (CKSAAP) is learned which can help to design a robust classifier. This eliminates the need for separately developing the classifier and the feature extraction modules, and a stand-alone model effectively learns the distinguishing characteristics of classes on a lower dimensional feature space.
The rest of the article is organized as follows: the classification framework of the proposed method is presented in Section 2, followed by the extensive experimentation and discussion in Section 3, and the study is concluded in Section 4.
2 Methods
2.1 Evaluation Metrics
For proper evaluation of the proposed model, a number of standard performance metrics have been used. The most intuitive performance measure is accuracy; however, for a highly imbalanced dataset (which is the case here), accuracy is not reflective of true performance. Therefore, various evaluation parameters, such as sensitivity, specificity, and Matthew’s correlation coefficient (MCC) are reported. Youden’s index and balanced accuracy are also considered to be important evaluation metrics for imbalanced data and are, therefore, extensively explored in this research.
2.2 Dataset
To design the proposed method, we used the benchmark dataset provided in Kandaswamy et al. (2013). The dataset consists of 445 ECM proteins and 3,327 non-ECM proteins. The 445 ECM proteins were curated from Swiss-Prot release 67 by first filtering 1103 ECM proteins from the pool of 17,233 metazoan-secreted protein sequences. Similarly, the negative dataset of 16,130 proteins were curated from secretory proteins that are annotated as non-ECM. Later, 445 ECM and 4,187 non-ECM nonhomologous sequences were further filtered out with the help of a clustering method (Li et al., 2001) by removing the sequences which showed 70% or higher similarity.
2.3 Feature Extraction
2.3.1 Composition of K-Spaced Amino Acid Pairs
One of the fundamental steps in designing a machine learning approach is the transformation of protein sequences to a numerical format. Several methods of this transformation exist and the resultant encoded vectors of the sequences are treated as the features. The common approach practiced by several researchers is to acquire various features of the same sequence by employing different encoding schemes, and their combination is utilized for training the machine learning algorithm. This laborious approach has resulted in the performance enhancement of some classifiers (Yu and Lu, 2011; Xiaowei et al., 2012; Yang et al., 2015; Xiao et al., 2016); however, some recent studies show that utilizing a single expedient-encoding scheme such as CKSAAP, which captures both short- and long-range interaction information between residues along the sequence, can result in an equally improved classification performance (Ju and Wang, 2018; Chen et al., 2019; Usman and Lee, 2019).
The CKSAAP scheme works on the simple principle of counting the occurrence frequencies of k-spaced amino acid pairs in the protein sequence. Each k-spaced amino acid pair represents the residue pair separated by any arbitrary number k (j = 0, 1, 2 … k) of amino acid residues. For k = 0, the encoding is similar to the DPC, in which protein sequence of 20 types of amino acids yields a feature vector of (20 × 20) = 400 types of amino acid pairs (i.e., AA, AC, AD, … YY)400. In earlier studies it has been suggested that the DPC and higher-order peptide features can be used to design a robust protein sequence classifier (Kandaswamy et al., 2011; Khan et al., 2018; Pratiwi et al., 2017). From Figure 1, it can be seen that for higher values of k, substantial neighborhood information is gathered for large peptide pairs. For instance k = 2, three feature segments, each having a length of 400, are obtained. These are then concatenated to get the final feature vector of length (k + 1) × 400. The graphical representation of the CKSAAP feature vector obtained with k = 2 has been depicted in Figure 1.
FIGURE 1. CKSAAP feature extraction mechanism for k = 2. Extracted from Usman et al. (2020).
This efficient method of encoding has, therefore, been favored by a number of researchers in various applications of computational biology including the prediction of anticancer peptides (Li et al., 2020), DNA, and several other binding sites (Ju and Wang, 2020; Lyu et al., 2020). Many adaptations of CKSAAP encoding scheme have utilized only the features generated by a single k value. In this research, we aim to find the optimal value of k by analyzing different combinations of the features generated by CKSAAP, and details are presented in Subsection 3.1.
2.4 Latent Space Learning for ECM Classification
Feature representation ability of the CKSAAP improves with large values of the parameter k, which is expected to result in a more robust model (Park et al., 2020b; Usman and Lee, 2019; Wu et al., 2019; Chen et al., 2017). However, the model utilizing a large number of features is susceptible to noise, resulting in a degraded performance. Furthermore, training the model on a large number of features not only results in an increased training time and complexity but is also prone to overfitting. To which end, feature selection/engineering, which involves the selection of most significant features, has to be employed. Feature selection techniques are broadly categorized into two types: 1) supervised methods, which remove the irrelevant features based on a target variable, and 2) unsupervised methods, which use correlation techniques to remove redundant information. A number of methods for feature selection have been proposed in the literature, including minimum redundancy maximum relevance (mRMR) (Peng et al., 2005), student’s t test (Student, 1908), info-gain (Mitchell et al., 1997), and generalized variant of strictly standardized mean difference (GSSMD) (Park et al., 2020a). Another useful method is to map the original data into a lower-order dimensional space through some transformation function. The eigen-space transformation or the principal component analysis method (PCA) (Jolliffe, 1986) is considered to be the benchmark method in this context. Other approaches such as an independent component analysis (ICA) (Comon, 1994), a kernel principal component analysis (KPCA) (Schölkopf et al., 1998), uniform manifold approximation and projection (UMAP) (McInnes et al., 2018), and t-distributed stochastic neighbor embedding (t-SNE) (Van der Maaten and Hinton, 2008) are also being successfully used to deal with the curse of dimensionality.
Most of the techniques mentioned above are unsupervised in nature. To address this issue, we propose to use a novel approach called a deep latent space encoding (DeepLSE) classifier for the latent space encoding based on an auto-encoder. Latent space refers to the representation of compressed data in which similar points would be in a close group, as shown in Figure 5. Similar samples tend to have common significance, which can be packaged into the latent space representation of the raw data. Thus, as the dimensions are reduced, the redundant information from the input samples is removed, leaving only the most important features of the data. In other words, the method can learn a compact representation of feature space and remove the noisy or potentially confusing information which is good for both the classification and reconstruction tasks. This ensures that the encoded features truly represent the sample information. The DeepLSE method has been found to be an impressive method for the feature space reduction and has outperformed other approaches in relatively similar tasks such as AFP-LSE (Usman et al., 2020) and E3-targetPred (Park et al., 2020b). The architecture of the proposed method is depicted in Figure 2 named as ECM-LSE.
FIGURE 2. Proposed DeepLSE architecture for ECM classification. The model comprises of an encoder, a decoder, and a classifier module. The encoder consists of an input layer and two hidden layers that embed input features to latent variables (LVs). The decoder architecture is mirror symmetry of the encoder which uses LVs as its input and generates the decoded output. The classification module uses latent space features as its input and four layers of fully connected neurons. Each hidden layer has 10 neurons, except for the last layer which produces a one-hot–encoded output of ECM/non-ECM class.
2.4.1 Network Specifications
The architecture of the proposed ECM-LSE network is composed of two modules: 1) an auto-encoder module and 2) a classification module.
2.4.1.1 Auto-Encoder Module
The auto-encoder is a type of neural network that can act as an identity function. It is used to find the representation of the input signal in a reduced dimensional space, known as the latent space. The principle of latent space–based representation is an assumption that the input signal has a low-rank. The auto-encoder network has a decoder that tries to regenerate the input from the latent space variables. During the training of an auto-encoder, the model is forced to become an identity function. Due to which only the relevant features of the data are learned in a compressed representation. This compressed representation has sufficient information for accurate reconstruction of the original input signal. The number of hidden layers and the number of neurons in each layer of the encoder and decoder are varied to obtain reasonable performance. In this research, the encoder and decoder are composed of three layers each, including two hidden layers. The number of neurons in the input layer of the encoder is equal to the length of the attribute vector. The number of neurons in the first and second hidden layers is set to be 50 and 10, respectively. The decoder is a mirror symmetry of the encoder. The number of neurons in the output layer of the decoder is equal to the length of the attribute vector. The number of neurons in the latent space is systematically altered to obtain the best performance for which we designed an ablation study discussed in Section 3.1. All hidden layers of the auto-encoder module are equipped with batch normalization, 30% dropout, and a rectified linear unit (ReLU) activation function. The latent space layer uses sigmoid activation function without any batch normalization and dropout.
2.4.1.2 Classification Module
The output of the encoder module (latent variables) is used as an input to the classification module. The classifier module shown in Figure 2 consists of four layers (three hidden and one output layer). All hidden layers consist of 10 neurons and a ReLU activation function. The last layer consists of two neurons representing the positive (ECM) and the negative (non-ECM) classes. For decision making, softmax activation function was used at the output layer.
3 Results
To develop a neural network model, the benchmark dataset was divided into the train, validation, and test datasets. For training, we formed a dataset consisting of 540 samples with equal number of ECMs and non-ECM protein samples. These were randomly selected from the pool of 445 ECMs and 3,327 non-ECMs, since the available dataset is very small, and it is highly likely that the model would suffer from the overfitting problem. To avoid such situation, we employed regularization techniques such as early stopping, dropout, batch normalization, and DeepLSE-based feature encoding. Furthermore, the validation dataset was also used with the aim of designing a generalized classifier module. The validation dataset consists of 30 ECMs and 810 non-ECMs randomly selected from the remaining 175 ECMs and 3,182 non-ECMs, respectively. The remaining 145 and 2,247 samples of ECMs and non-ECMs were used in the test dataset. Several model configurations on the basis of the latent space size (LVs) and the CKSAAP gap value k were evaluated. For each choice of model configuration, the process of model training was repeated 20 times and mean and standard deviations of performance statistics were reported. In each trial, the weights and bias of the model were randomly initialized. Also, each trial utilized randomly configured subsets from the training, validation, and test dataset. The validation process assisted toward the filtration of the overfitted models, that is, only the models with 75% or higher validation balanced accuracy was selected.
3.1 Ablation Study
The workflow of the proposed study is aimed to obtain the best classification model based on two variables, that is, the gap between the two amino acid pairs and the number of units in the latent space LVs. An ablation study has been designed to acquire models with varying number of aforementioned variables and is depicted in Figure 3 (a). The samples are distributed into training, validation, and test datasets as discussed in the Subsection 2.2 and are encoded with incrementing values of k from 0 to 10. The resultant features are used to train the model with incrementing values of the latent space variables ranging from 2 to 9. As discussed earlier, for each configuration, 20 independent trials are performed and the mean results are computed. A consistent procedure is repeated for all 1,760 trials of the 88 unique model configurations. The model with the best average results is finally selected as the base model to perform prediction and is named as ECM-LSE. In Table. 1, the average results of the balanced accuracy have been reported. It can be observed that the model with values of gap k = 8 and latent variables LV = 7, accounts for the best. The results for the rest of the evaluation parameters are illustrated in the form of surface graphs in Figure 4.
FIGURE 4. Performance statistics surfaces for: (A) accuracy, (B) MCC, (C) balanced-accuracy, (D) Youden’s Index, (E) F1-Score, and (F) mean squared error (MSE) in dB.
3.2 Comparison With Contemporary Approaches
The performance of the proposed model is compared to the benchmark approaches and the findings are reported in Table 2. For a fair comparison, only the best reported results of the respective approaches are presented. The performance of the proposed ECM-LSE is compared with the contemporary methods including EcmPred (Kandaswamy et al., 2013), a sparse learning approach for the prediction of ECM (ECMSRC) (Naseem et al., 2017), and PECM (Zhang et al., 2014). In particular, the reported sensitivity, specificity, MCC, Youden’s index, and accuracy on the benchmark dataset of EcmPred (Kandaswamy et al., 2013) are compared.
TABLE 2. Comparison of the proposed ECM-LSE algorithm with the benchmark machine learning approaches on the test dataset.
The results clearly show that the proposed method has better balanced accuracy as compared to the contemporary approaches. In particular, the proposed ECM-LSE method achieves the highest sensitivity of 84.14% outperforming the best competitor (PECM) by a margin of 10.91%. The specificity value achieved by the proposed ECM-LSE also compares favorably with other methods, which confirms the balanced unbiased learning effect. It is noteworthy to point out that the accuracy metric cannot provide true fitness of the models given the skewed distribution of test dataset toward the negative (non-ECM) class. Any model with all negative predictions can achieve
Furthermore, unlike contemporary methods where handcrafted embedding schemes are utilized for separately developing the classifier and the feature extraction modules, the proposed ECM-LSE method learns directly from the original feature space. The LSE encoding effectively learns the distinguishing characteristics of classes in a lower dimensional feature space and allows the visualization of proteins sequences. This aspect of ECM-LSE is further explained in Section 3.4.
3.3 Verification on Experimentally Verified Human ECM Proteins
To verify the practical usefulness of our method, herein, we perform the validation of our method on experimentally verified ECM proteins. In particular, we collected 20 experimentally verified human ECM proteins from UniProt (Consortium, 2018). The collected sequences were not present in the positive or negative datasets of ECM-LSE. The criteria for the selection were based on the clear experimental evidence in the literature for the given sequence entry. We evaluated the EcmPred (Kandaswamy et al., 2013), ECMSRC (Naseem et al., 2017), PECM (Zhang et al., 2014), and ECM-LSE methods. As shown in Table 3, ECM-LSE (k = 8 and LV = 7) correctly identified 19 proteins as extracellular matrix proteins, whereas PECM, ECMSRC, and EcmPred identified 18, 16, and 15 proteins, respectively. It is noteworthy to point out that the models were trained on ECM proteins from metazoans; therefore, the superior performance of the proposed ECM-LSE on proteins from a completely different organism suggests that it can be effectively utilized for the annotation of unknown proteins.
TABLE 3. Prediction results for 20 experimentally verified ECM proteins. “✔” indicates correctly identification while “✗” represents an incorrect identification.
3.4 Discussion
For typical classification problems such as lysine acetylation site prediction in proteins (Wu et al., 2019) or the identification of protein–protein binding sites (Fernandez-Recio et al., 2005), a large number of positive and negative samples are usually available in the datasets. Therefore, the problem of class imbalance or intra-class variation is not a major concern (Johnson and Khoshgoftaar, 2019). However, the limited availability of ECM samples results in an imbalanced dataset, resulting in an ill-posed problem. A number of approaches, including sample rescaling, have been proposed in the literature to tackle the imbalanced data problem (Xiao et al., 2016; Kabir et al., 2018). Classifiers based on these rescaling techniques tend to behave well; however, the generalization of the method is compromised. Furthermore, the comparison of methods using rescaled samples with the methods using a standard dataset is not reasonable. In the proposed study, we utilize a standard dataset and develop a method that effectively discriminates the ECM proteins from non-ECM. This is achieved through the latent space learning of the CKSAAP features. For better understanding, we compare the t-SNE projection of the CKSAAP features with the proposed latent space in Figure 5.
FIGURE 5. Feature embedding (A) using t-SNE (Van der Maaten and Hinton, 2008) projection of the original feature space and (B) using the proposed latent space encoding (ECM-LSE) method.
For visualization purposes, the data were projected on two dimensions using t-SNE (Van der Maaten and Hinton, 2008) projection of the original feature space and two variable latent spaces in the case of ECM-LSE. In the t-SNE projection shown in Figure 5A, it can be observed that both ECMs and non-ECMs appear in an overlapping fashion, suggesting that the development of the ECM classifier using original feature space is an arduous task. As shown in Figure 5B, the proposed latent space encoding (ECM-LSE) presents superior learning capabilities and maps the ECMs and non-ECMs in separate regions in contrast to the unsupervised subspace learning method of t-SNE (Van der Maaten and Hinton, 2008).
The proposed method, as shown in Figure 5B, tends to form distinguishable clusters of ECM and non-ECM proteins. Although some overlap can be observed in the projection of the proposed method, it is still remarkably better than that of the t-SNE, and since the projection is shown for two latent variables only, the actual model with seven latent variables is expected to mitigate the overlap to a greater extent. These projections are also helpful in understanding the working principle of the proposed method and the motivation for the development of nonlinear auto-encoded learning of latent space.
The proposed hybrid approach presents a hybrid design with capabilities of efficient feature selection and classification of ECM proteins. The latent space dynamically reduces the dimension of the feature space and retains only the relevant information sufficient to efficiently distinguish ECM from non-ECM samples. Although, the proposed method can predict ECM from different organisms, it is not a replacement for gold standard wet lab–based testing. Furthermore, due to the scarcity of available ECM proteins the model may show biased performance in favor of already explored ECM and finding novel proteins may require the fusion of additional information. However, efforts have been made to avoid overfitting in order to seek the generalization property of the model by deploying dropout and batch normalization techniques. Further enhancements to the ECM prediction task where scarcity of the positive samples persists can be made by applying a transfer learning approach, where a large scale model is trained on a closely related dataset and is further fine-tuned for ECM samples. The Python implementation of the proposed algorithm has been made public, and interested users can utilize the algorithm for their problem of interest. The algorithm is available at (https://github.com/Shujaat123/ECM-LSE/blob/master/ECM_LSE_Online.ipynb). In the future, we aim to explore the efficacy of the auto-encoder–based classifiers on other bioinformatics problems.
4 Conclusion
ECM is a complex meshwork of cross-linked proteins responsible for the architectural support of cells and contributes to the functionality of the living tissue. They also contribute toward the formation of the cancer stem cells; therefore, their study and classification from non-ECMs proteins is of prime importance. A reliable prediction method can not only help understand various abnormalities associated with several cancer types but will also assist in diagnostic research. Conventional experimental-based methods are considered gold standards for this task; however, they are extremely time consuming and scanning a large number of proteins is practically infeasible. In this research, we designed a latent space learning method for the classification of ECM proteins. The proposed method can be used as a reliable prediction model. An important feature of the proposed method is its latent space-based projections through which protein sequences can be visualized in filtered and reduced dimensions, which is extremely helpful in finding useful clusters. The proposed method has been tested on a benchmark dataset and results of widely used performance metrics are reported. In particular, we report a balanced test accuracy of 86.45% with 0.71 Youden’s index and 0.39 MCC (with k = 8 and LV = 7). Additionally, the model performance is verified on completely unseen experimentally verified ECM proteins and shown to achieve highest prediction score.
Data Availability Statement
Publicly available datasets were analyzed in this study. This data can be found here: https://github.com/Shujaat123/ECM-LSE.
Author Contributions
All authors read the final manuscript and validated the results. Specific individual contributions are as follows: UA-S: supervision. MU: visualization and writing—original draft preparation. IN: writing—reviewing and editing. MN: project administration. AJ: writing—reviewing and editing. MA: funding acquisition. HA: funding acquisition. SK: conceptualization, visualization, methodology, and investigation.
Funding
This research work was funded by the Institutional Fund Project under grant no. IFPHI-139-135-2020. Therefore, the authors gratefully acknowledge technical and financial support from the Ministry of Education and King Abdulaziz University, DSR, Jeddah, Saudi Arabia.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors, and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
The author would like to thank the UniProtKB and NCBI community for providing public database of protein sequences.
References
Asim, M. N., Ibrahim, M. A., Imran Malik, M., Dengel, A., and Ahmed, S. (2021). Advances in Computational Methodologies for Classification and Sub-cellular Locality Prediction of Non-coding Rnas. Ijms 22, 8719. doi:10.3390/ijms22168719
Chen, J., Zhao, J., Yang, S., Chen, Z., and Zhang, Z. (2019). Prediction of Protein Ubiquitination Sites in Arabidopsis Thaliana. Cbio 14, 614–620. doi:10.2174/1574893614666190311141647
Chen, Q.-Y., Tang, J., and Du, P.-F. (2017). Predicting Protein Lysine Phosphoglycerylation Sites by Hybridizing many Sequence Based Features. Mol. Biosyst. 13, 874–882. doi:10.1039/c6mb00875e
Chicco, D., and Jurman, G. (2020). The Advantages of the matthews Correlation Coefficient (Mcc) over F1 Score and Accuracy in Binary Classification Evaluation. BMC genomics 21, 6–13. doi:10.1186/s12864-019-6413-7
Chou, K.-C. (2011). Some Remarks on Protein Attribute Prediction and Pseudo Amino Acid Composition. J. Theor. Biol. 273, 236–247. doi:10.1016/j.jtbi.2010.12.024
Chou, K.-C., Wu, Z.-C., and Xiao, X. (2012). iLoc-Hum: Using the Accumulation-Label Scale to Predict Subcellular Locations of Human Proteins with Both Single and Multiple Sites. Mol. Biosyst. 8, 629–641. doi:10.1039/c1mb05420a
Comon, P. (1994). Independent Component Analysis, a New Concept. Signal. Processing 36, 287–314. doi:10.1016/0165-1684(94)90029-9
Consortium, T. U. (2018). UniProt: a Worldwide Hub of Protein Knowledge. Nucleic Acids Res. 47, D506–D515. doi:10.1093/nar/gky1049
Ding, X., Zou, Z., and Brooks Iii, C. L. (2019). Deciphering Protein Evolution and Fitness Landscapes with Latent Space Models. Nat. Commun. 10, 5644–5657. doi:10.1038/s41467-019-13633-0
Dyrløv Bendtsen, J., Nielsen, H., von Heijne, G., and Brunak, S. (2004). Improved Prediction of Signal Peptides: SignalP 3.0. J. Mol. Biol. 340, 783–795. doi:10.1016/j.jmb.2004.05.028
Endo, Y., Ishiwata-Endo, H., and Yamada, K. M. (2012). Extracellular Matrix Protein Anosmin Promotes Neural Crest Formation and Regulates Fgf, Bmp, and Wnt Activities. Developmental Cel. 23, 305–316. doi:10.1016/j.devcel.2012.07.006
Fernandez-Recio, J., Totrov, M., Skorodumov, C., and Abagyan, R. (2005). Optimal Docking Area: a New Method for Predicting Protein-Protein Interaction Sites. Proteins 58, 134–143. doi:10.1002/prot.20285
Gonzalez-Pujana, A., Santos-Vizcaino, E., García-Hernando, M., Hernaez-Estrada, B., M. de Pancorbo, M., Benito-Lopez, F., et al. (2019). Extracellular Matrix Protein Microarray-Based Biosensor with Single Cell Resolution: Integrin Profiling and Characterization of Cell-Biomaterial Interactions. Sensors Actuators B: Chem. 299, 126954. doi:10.1016/j.snb.2019.126954
Hanchuan Peng, H., Fuhui Long, F., and Ding, C. (2005). Feature Selection Based on Mutual Information Criteria of max-dependency, max-relevance, and Min-Redundancy. IEEE Trans. Pattern Anal. Machine Intell. 27, 1226–1238. doi:10.1109/tpami.2005.159
Hiraki, H. L., Matera, D. L., Rose, M. J., Kent, R. N., Todd, C. W., Stout, M. E., et al. (2021). Magnetic Alignment of Electrospun Fiber Segments within a Hydrogel Composite Guides Cell Spreading and Migration Phenotype Switching. Front. Bioeng. Biotechnol. 9, 679165. doi:10.3389/fbioe.2021.679165
Horton, P., Park, K.-J., Obayashi, T., Fujita, N., Harada, H., Adams-Collier, C. J., et al. (2007). WoLF PSORT: Protein Localization Predictor. Nucleic Acids Res. 35, W585–W587. doi:10.1093/nar/gkm259
Horton, P., Park, K., Obayashi, T., and Nakai, K. (2006). Protein Subcellular Localisation Prediction with WoLF PSORT. APBC 35, 39–48.
Hou, Z., Yang, Y., Li, H., Wong, K.-c., and Li, X. (2021). Ideepsubmito: Identification of Protein Submitochondrial Localization with Deep Learning. Brief. Bioinform., bbab288. doi:10.1093/bib/bbab288
Hu, J., Van den Steen, P. E., Sang, Q.-X. A., and Opdenakker, G. (2007). Matrix Metalloproteinase Inhibitors as Therapy for Inflammatory and Vascular Diseases. Nat. Rev. Drug Discov. 6, 480–498. doi:10.1038/nrd2308
Johnson, J. M., and Khoshgoftaar, T. M. (2019). Survey on Deep Learning with Class Imbalance. J. Big Data 6, 27. doi:10.1186/s40537-019-0192-5
Jolliffe, I. T. (1986). Principal Components in Regression Analysis. Principal Component Analysis. Springer, 129–155. doi:10.1007/978-1-4757-1904-8_8
Ju, Z., and Wang, S.-Y. (2018). Prediction of Citrullination Sites by Incorporating K-Spaced Amino Acid Pairs into Chou's General Pseudo Amino Acid Composition. Gene 664, 78–83. doi:10.1016/j.gene.2018.04.055
Ju, Z., and Wang, S.-Y. (2020). Prediction of Lysine Formylation Sites Using the Composition of K-Spaced Amino Acid Pairs via Chou's 5-steps Rule and General Pseudo Components. Genomics 112, 859–866. doi:10.1016/j.ygeno.2019.05.027
Jung, J., Ryu, T., Hwang, Y., Lee, E., and Lee, D. (2010). Prediction of Extracellular Matrix Proteins Based on Distinctive Sequence and Domain Characteristics. J. Comput. Biol. 17, 97–105. doi:10.1089/cmb.2008.0236
Kabir, M., Ahmad, S., Iqbal, M., Khan Swati, Z. N., Liu, Z., and Yu, D.-J. (2018). Improving Prediction of Extracellular Matrix Proteins Using Evolutionary Information via a Grey System Model and Asymmetric Under-sampling Technique. Chemometrics Intell. Lab. Syst. 174, 22–32. doi:10.1016/j.chemolab.2018.01.004
Kandaswamy, K. K., Chou, K.-C., Martinetz, T., Möller, S., Suganthan, P. N., Sridharan, S., et al. (2011). AFP-pred: A Random forest Approach for Predicting Antifreeze Proteins from Sequence-Derived Properties. J. Theor. Biol. 270, 56–62. doi:10.1016/j.jtbi.2010.10.037
Kandaswamy, K. K., Pugalenthi, G., Hartmann, E., Kalies, K.-U., Möller, S., Suganthan, P. N., et al. (2010). SPRED: A Machine Learning Approach for the Identification of Classical and Non-classical Secretory Proteins in Mammalian Genomes. Biochem. Biophysical Res. Commun. 391, 1306–1311. doi:10.1016/j.bbrc.2009.12.019
Kandaswamy, K. K., Pugalenthi, G., Kalies, K.-U., Hartmann, E., and Martinetz, T. (2013). EcmPred: Prediction of Extracellular Matrix Proteins Based on Random forest with Maximum Relevance Minimum Redundancy Feature Selection. J. Theor. Biol. 317, 377–383. doi:10.1016/j.jtbi.2012.10.015
Karagöz, Z., Geuens, T., LaPointe, V. L., van Griensven, M., and Carlier, A. (2021). Win, Lose, or Tie: Mathematical Modeling of Ligand Competition at the Cell–Extracellular Matrix Interface. Front. Bioeng. Biotechnol. 9, 340. doi:10.3389/fbioe.2021.657244
Khan, S., Naseem, I., Togneri, R., and Bennamoun, M. (2018). Rafp-pred: Robust Prediction of Antifreeze Proteins Using Localized Analysis of N-Peptide Compositions. Ieee/acm Trans. Comput. Biol. Bioinf. 15, 244–250. doi:10.1109/tcbb.2016.2617337
Kim, S.-H., Turnbull, J., and Guimond, S. (2011). Extracellular Matrix and Cell Signalling: the Dynamic Cooperation of Integrin, Proteoglycan and Growth Factor Receptor. J. Endocrinol. 209, 139–151. doi:10.1530/joe-10-0377
Kizawa, H., Kou, I., Iida, A., Sudo, A., Miyamoto, Y., Fukuda, A., et al. (2005). An Aspartic Acid Repeat Polymorphism in Asporin Inhibits Chondrogenesis and Increases Susceptibility to Osteoarthritis. Nat. Genet. 37, 138–144. doi:10.1038/ng1496
Klavert, J., and van der Eerden, B. C. (2021). Fibronectin in Fracture Healing: Biological Mechanisms and Regenerative Avenues. Front. Bioeng. Biotechnol. 9, 274. doi:10.3389/fbioe.2021.663357
Klee, E. W., and Sosa, C. P. (2007). Computational Classification of Classically Secreted Proteins. Drug Discov. Today 12, 234–240. doi:10.1016/j.drudis.2007.01.008
Lewis, D. D., Villarreal, F. D., Wu, F., and Tan, C. (2014). Synthetic Biology outside the Cell: Linking Computational Tools to Cell-free Systems. Front. Bioeng. Biotechnol. 2, 66. doi:10.3389/fbioe.2014.00066
Li, Q., Zhou, W., Wang, D., Wang, S., and Li, Q. (2020). Prediction of Anticancer Peptides Using a Low-Dimensional Feature Model. Front. Bioeng. Biotechnol. 8, 892. doi:10.3389/fbioe.2020.00892
Li, W., Jaroszewski, L., and Godzik, A. (2001). Clustering of Highly Homologous Sequences to Reduce the Size of Large Protein Databases. Bioinformatics 17, 282–283. doi:10.1093/bioinformatics/17.3.282
Lyu, X., Li, S., Jiang, C., He, N., Chen, Z., Zou, Y., et al. (2020). Deepcso: a Deep-Learning Network Approach to Predicting Cysteine S-Sulphenylation Sites. Front Cel Dev Biol. 8, 594587. doi:10.3389/fcell.2020.594587
Lyu, Z., Wang, Z., Luo, F., Shuai, J., and Huang, Y. (2021). Protein Secondary Structure Prediction with a Reductive Deep Learning Method. Front. Bioeng. Biotechnol. 9, 687426. doi:10.3389/fbioe.2021.687426
Ma, F., Tremmel, D. M., Li, Z., Lietz, C. B., Sackett, S. D., Odorico, J. S., et al. (2019). In Depth Quantification of Extracellular Matrix Proteins from Human Pancreas. J. Proteome Res. 18, 3156–3165. doi:10.1021/acs.jproteome.9b00241
Mathews, S., Bhonde, R., Gupta, P. K., and Totey, S. (2012). Extracellular Matrix Protein Mediated Regulation of the Osteoblast Differentiation of Bone Marrow Derived Human Mesenchymal Stem Cells. Differentiation 84, 185–192. doi:10.1016/j.diff.2012.05.001
McInnes, L., Healy, J., and Melville, J. (2018). Umap: Uniform Manifold Approximation and Projection for Dimension Reduction. arXiv preprint arXiv:1802.03426
Mikolov, T., Sutskever, I., Chen, K., Corrado, G., and Dean, J. (2013). Distributed Representations of Words and Phrases and Their Compositionality. arXiv preprint arXiv:1310.4546.
Naseem, I., Khan, S., Togneri, R., and Bennamoun, M. (2017). Ecmsrc: A Sparse Learning Approach for the Prediction of Extracellular Matrix Proteins. Curr. Bioinformatics 12, 361–368. doi:10.2174/1574893611666151215213508
Olsen, T. H., Yesiltas, B., Marin, F. I., Pertseva, M., García-Moreno, P. J., Gregersen, S., et al. (2020). Anoxpepred: Using Deep Learning for the Prediction of Antioxidative Properties of Peptides. Sci. Rep. 10, 21471–21481. doi:10.1038/s41598-020-78319-w
Otzen, D. E., Dueholm, M. S., Najarzadeh, Z., Knowles, T. P. J., and Ruggeri, F. S. (2021). In Situ Sub‐Cellular Identification of Functional Amyloids in Bacteria and Archaea by Infrared Nanospectroscopy. Small Methods 5, 2001002. doi:10.1002/smtd.202001002
Park, S., Khan, S., Moinuddin, M., and Al-Saggaf, U. M. (2020a). Gssmd: A New Standardized Effect Size Measure to Improve Robustness and Interpretability in Biological Applications. In 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, South Korea, December 16-19, 2020 (IEEE), 1096–1099. doi:10.1109/bibm49941.2020.9313582
Park, S., Khan, S., and Wahab, A. (2020b). E3-targetpred: Prediction of e3-target proteins using deep latent space encoding. arXiv preprint arXiv:2007.12073
Pratiwi, R., Malik, A. A., Schaduangrat, N., Prachayasittikul, V., Wikberg, J. E., Nantasenamat, C., et al. (2017). Cryoprotect: a Web Server for Classifying Antifreeze Proteins from Nonantifreeze Proteins. J. Chem. 2017. doi:10.1155/2017/9861752
Ras-Carmona, A., Gomez-Perosanz, M., and Reche, P. A. (2021). Prediction of Unconventional Protein Secretion by Exosomes. BMC bioinformatics 22, 333–346. doi:10.1186/s12859-021-04219-z
Schölkopf, B., Smola, A., and Müller, K.-R. (1998). Nonlinear Component Analysis as a Kernel Eigenvalue Problem. Neural Comput. 10, 1299–1319. doi:10.1162/089976698300017467
Shen, F., Cai, W., Gan, X., Feng, J., Chen, Z., Guo, M., et al. (2021). Prediction of Genetic Factors of Hyperthyroidism Based on Gene Interaction Network. Front. Cel Developmental Biol., 1668. doi:10.3389/fcell.2021.700355
Usman, M., Khan, S., and Lee, J. A. (2020). Afp-lse: Antifreeze Proteins Prediction Using Latent Space Encoding of Composition of K-Spaced Amino Acid Pairs. Sci. Rep. 10, 7197–7210. doi:10.1038/s41598-020-63259-2
Usman, M., and Lee, J. A. (2019). Afp-cksaap: Prediction of Antifreeze Proteins Using Composition of K-Spaced Amino Acid Pairs with Deep Neural Network. In 2019 IEEE 19th International Conference on Bioinformatics and Bioengineering (BIBE), Athens, Greece, October 28-30, 2019, (IEEE), 38–43. doi:10.1109/bibe.2019.00016
Wang, Y., Wang, P., Guo, Y., Huang, S., Chen, Y., and Xu, L. (2021). Prpred: A Predictor to Identify Plant Resistance Proteins by Incorporating K-Spaced Amino Acid (Group) Pairs. Front. Bioeng. Biotechnol. 8, 1593. doi:10.3389/fbioe.2020.645520
Wu, M., Yang, Y., Wang, H., and Xu, Y. (2019). A Deep Learning Method to More Accurately Recall Known Lysine Acetylation Sites. BMC bioinformatics 20, 49. doi:10.1186/s12859-019-2632-9
Wu, Z.-C., Xiao, X., and Chou, K.-C. (2012). iLoc-Gpos: A Multi-Layer Classifier for Predicting the Subcellular Localization of Singleplex and Multiplex Gram-Positive Bacterial Proteins. Ppl 19, 4–14. doi:10.2174/092986612798472839
Wu, Z.-C., Xiao, X., and Chou, K.-C. (2011). iLoc-Plant: A Multi-Label Classifier for Predicting the Subcellular Localization of Plant Proteins with Both Single and Multiple Sites. Mol. Biosyst. 7, 3287–3297. doi:10.1039/c1mb05232b
Xiao, X., Hui, M., and Liu, Z. (2016). Iafp-Ense: an Ensemble Classifier for Identifying Antifreeze Protein by Incorporating Grey Model and Pssm into Pseaac. J. Membr. Biol. 249, 845–854. doi:10.1007/s00232-016-9935-9
Xiao, X., Wu, Z.-C., and Chou, K.-C. (2011). iLoc-Virus: A Multi-Label Learning Classifier for Identifying the Subcellular Localization of Virus Proteins with Both Single and Multiple Sites. J. Theor. Biol. 284, 42–51. doi:10.1016/j.jtbi.2011.06.005
Xiaowei, Z., Zhiqiang, M., and Minghao, Y. (2012). Using Support Vector Machine and Evolutionary Profiles to Predict Antifreeze Protein Sequences. Int. J. Mol. Sci. 13, 2196–2207.
Yang, R., Zhang, C., Gao, R., and Zhang, L. (2015). An Effective Antifreeze Protein Predictor with Ensemble Classifiers and Comprehensive Sequence Descriptors. Ijms 16, 21191–21214. doi:10.3390/ijms160921191
Yu, C.-S., and Lu, C.-H. (2011). Identification of Antifreeze Proteins and Their Functional Residues by Support Vector Machine and Genetic Algorithms Based on N-Peptide Compositions. PloS one 6, e20445. doi:10.1371/journal.pone.0020445
Zemouri, R. (2020). Semi-supervised Adversarial Variational Autoencoder. Make 2, 361–378. doi:10.3390/make2030020
Zhang, D., Chen, H.-D., Zulfiqar, H., Yuan, S.-S., Huang, Q.-L., Zhang, Z.-Y., et al. (2021). Iblp: An Xgboost-Based Predictor for Identifying Bioluminescent Proteins. Comput. Math. Methods Med. 2021, 6664362. doi:10.1155/2021/6664362
Zhang, J., Sun, P., Zhao, X., and Ma, Z. (2014). PECM: Prediction of Extracellular Matrix Proteins Using the Concept of Chou's Pseudo Amino Acid Composition. J. Theor. Biol. 363, 412–418. doi:10.1016/j.jtbi.2014.08.002
Keywords: extracellular matrix (ECM), auto-encoder, composition of k-spaced amino acid pair (CKSAAP), latent space learning, neural network, classification, amino acid composition (AAC)
Citation: Al-Saggaf UM, Usman M, Naseem I, Moinuddin M, Jiman AA, Alsaggaf MU, Alshoubaki HK and Khan S (2021) ECM-LSE: Prediction of Extracellular Matrix Proteins Using Deep Latent Space Encoding of k-Spaced Amino Acid Pairs. Front. Bioeng. Biotechnol. 9:752658. doi: 10.3389/fbioe.2021.752658
Received: 03 August 2021; Accepted: 13 September 2021;
Published: 14 October 2021.
Edited by:
Zhiguang Zhu, Tianjin Institute of Industrial Biotechnology, (CAS), ChinaReviewed by:
Mario Andrea Marchisio, Tianjin University, ChinaGe Qu, Tianjin Institute of Industrial Biotechnology, (CAS), China
Copyright © 2021 Al-Saggaf, Usman, Naseem, Moinuddin, Jiman, Alsaggaf, Alshoubaki and Khan. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Shujaat Khan , c2h1amFhdEBrYWlzdC5hYy5rcg==