- 1Center for Applied Scientific Computing, Lawrence Livermore National Laboratory, Livermore, CA, United States
- 2Biosciences and Biotechnology Division, Lawrence Livermore National Laboratory, Livermore, CA, United States
- 3Global Security Computing Applications Division, Lawrence Livermore National Laboratory, Livermore, CA, United States
- 4Computational Engineering Division, Lawrence Livermore National Laboratory, Livermore, CA, United States
Introduction: Recent advances in 3D structure-based deep learning approaches demonstrate improved accuracy in predicting protein-ligand binding affinity in drug discovery. These methods complement physics-based computational modeling such as molecular docking for virtual high-throughput screening. Despite recent advances and improved predictive performance, most methods in this category primarily rely on utilizing co-crystal complex structures and experimentally measured binding affinities as both input and output data for model training. Nevertheless, co-crystal complex structures are not readily available and the inaccurate predicted structures from molecular docking can degrade the accuracy of the machine learning methods.
Methods: We introduce a novel structure-based inference method utilizing multiple molecular docking poses for each complex entity. Our proposed method employs multi-instance learning with an attention network to predict binding affinity from a collection of docking poses.
Results: We validate our method using multiple datasets, including PDBbind and compounds targeting the main protease of SARS-CoV-2. The results demonstrate that our method leveraging docking poses is competitive with other state-of-the-art inference models that depend on co-crystal structures.
Discussion: This method offers binding affinity prediction without requiring co-crystal structures, thereby increasing its applicability to protein targets lacking such data.
1 Introduction
Given the vast number of drug-like chemical compounds and the rapid growth of compound library databases, it is crucial to develop an effective tool for large-scale virtual high-throughput screening to identify hit molecules. To address this need, recent studies have employed various machine learning and deep learning-based inference models to enhance traditional computational modeling techniques such as molecular docking and molecular dynamics simulations for accurately predicting protein-ligand interactions and binding affinities.
One major application utilizing machine learning techniques in drug discovery involves the use of SMILES strings and their associated features such as molecular descriptors (RDKit, 2024; Moriwaki et al., 2018) and fingerprints (Rogers and Hahn, 2010), to predict compound properties and target-specific interactions. This includes predicting toxicity, solubility, membrane permeability, binding affinity, and other compound characteristics associated with the target receptors. The approaches in this category use a variety of machine learning methods, ranging from random forest algorithms to various neural network architectures (Minnich et al., 2020; Ramsundar et al., 2019).
Recent studies use advanced architectures and techniques with atomic graph representations such as graph neural networks (Duvenaud et al., 2015; Stärk et al., 2022a), graph transformers and attention mechanisms (Rong et al., 2020) in unsupervised or self-supervised settings. These approaches show the advantages of utilizing pretrained models on unlabeled large-volume compound databases. Furthermore, large language model-based approaches have been proposed to address pretrainng and finetuning with the unlabeled compound data (Chithrananda et al., 2020; Mazuz et al., 2023).
Another direction in leveraging deep learning techniques uses 3D atomic representations of the protein-ligand structures to predict the interaction and binding affinities. These structure-based methods use 3D convolutional neural networks (3DCNNs) utilizing voxelized atomic representations of co-crystal complex structures to predict binding affinities (Gomes et al., 2017; Ragoza et al., 2017; Jiménez et al., 2018; Zhang et al., 2019) and to predict protein-binding sites (Jiménez et al., 2017). Furthermore, graph neural networks (GNNs), graph convolutional networks (GCNs), and spatial graph convolutional neural networks (SGCNNs) have been introduced for binding affinity prediction. These models utilize graph representations of atomic data from co-crystal complex structures, capturing atoms and their connectivity through nodes and edges (Feinberg et al., 2018; 2019; Zhang et al., 2023). Recently, equivariant attention networks have received attention for representing 3D point cloud or graph data under rotations such as SE(3)-transformers (Fuchs et al., 2020). Satorras et al. developed light-weight E(n) equivariant graph neural networks (EGNNs) equivariant to rotations, translations, reflections and permutations without the need for higher-order representations (Satorras et al., 2021). Powers et al. proposed to use data-efficient E(3) equivariant networks with point cloud data to guide how to attach new molecular fragments in molecule optimization (Powers et al., 2022). Scantlebury et al. proposed a lightweight E(n)-equivariant graph neural network for machine learning-based scoring functions (Scantlebury et al., 2023).
An additional recent trend involves using hybrid neural network architectures or fusion networks to capture diverse input representations and interactions. Jones et al. proposed a fusion method to complement 3DCNN and SGCNN architectures to effectively capture both spatial information and atom connectivities (Jones et al., 2021). Jiang et al. proposed the use of two independent graph network modules to capture intra- and intermolecular interactions (Jiang et al., 2021). Kyro et al. proposed the use of hybrid attentions with 3DCNN and GCN (Kyro et al., 2023). Mqawass et al. proposed a fusion method of different input representations (Mqawass and Popov, 2024).
Most methods in this category rely on co-crystal complex structures of protein-ligand data to train the neural network models. However, crystal structures are not readily available in the early stages of most drug discovery applications. The PDBbind dataset (Su et al., 2019), a meticulously curated subset of the Protein Data Bank (PDB) (Burley et al., 2019) containing experimental binding affinity data, is virtually the sole dataset accessible for structure-based protein-ligand learning tasks. For that reason, we often utilize simulated docking poses of protein-ligand complex structures for structure-based deep learning inference models. The docking poses can be generated using molecular docking tools such as AutoDock VINA (Trott and Olson, 2010; Eberhardt et al., 2021), and GLIDE (Friesner et al., 2004; Halgren et al., 2004). Recent studies have proposed the use of deep neural network approaches to serve as docking scoring functions (McNutt et al., 2021), or to generate docking poses including DeepDock (Méndez-Lucio et al., 2021), EquiBind (Stärk et al., 2022b), TankBind (Lu et al., 2022), DiffDock (Corso et al., 2023), and UniMol (Zhou et al., 2023).
While generating docking poses is feasible and practical, training or evaluating deep neural network models on these poses results in inaccurate predictions, as the docking poses are considered “noisy” data. To identify inaccurate docking poses, Shim et al. proposed a pose classification approach using two different neural network methods, 3DCNN and point cloud network (PCN) (Shim et al., 2022). While the method filters out inaccurate docking poses; it requires an additional computational or inference method for binding affinity estimation. Another study developed a pose classification method without using crystal structures to improve the correlation between docking scores and experimental data (Shim et al., 2024). However, their method relies on the docking poses and their scoring functions. A recently proposed method leverages an E(3)-invariant graph neural network, combined with docking and binding affinity data, to address binding affinity prediction for kinase compounds (Backenköhler et al., 2024). Although this approach utilizes template docking to achieve more accurate poses, it does not explicitly address the distinction between accurate and inaccurate docking data, particularly in scenarios where a significant amount of inaccurate docking data is present.
In this paper, we present a novel inference method based on multi-instance learning (MIL) that utilizes a set of docking poses for each protein-ligand entity to predict binding affinity. Multi-instance learning is a type of weakly supervised learning used to predict a label for a group (bag) of instances, where the individual labels of the instances within the bag are not available (Carbonneau et al., 2018). In the standard MIL setup for classification, it is assumed that negative bags exclusively contain negative instances, while positive bags contain at least one positive instance. This technique offers flexibility by mitigating the potential problems associated with ambiguous and time-consuming labeling, compared to the traditional supervised learning. For more detail about MIL methods in various computer vision and machine learning applications, refer to the following review papers (Carbonneau et al., 2018; Fatima et al., 2023).
We leverage MIL-based regression to predict binding affinities when only molecular docking data is available, as there are no accessible co-crystal structures. Our method begins with graph neural networks, specifically SGCNN and EGNN, representing each ligand pose alongside its associated target receptor instance as a graph embedding. We then extract latent features from multiple pose instances per complex entity using these graph networks. Finally, we integrate these features into our pose-wise attention network, which functions as a MIL framework to predict binding affinities for each complex entity. We train and evaluate our models on molecular docking structures that may contain inaccurate poses. We demonstrate our method using the PDBbind and SARS-CoV-2 main protease datasets. The primary contribution of our method is as follows:
2 Methods
2.1 Problem definition
Most high-throughput screening and 3D structure-based neural network approaches utilize co-crystal complex structures or the “best” poses with the highest docking scores (lowest binding energy). Instead, we consider all generated poses with the target receptor as a bag for each complex structure entity. We formulate this problem as multi-instance learning (MIL) regression to predict a single real-value label for a bag (Dooly et al., 2002). Each bag represents a single protein-ligand structure entity consisting of multiple docking poses,
2.2 Input data and featurization
For input ligand docking poses and protein receptors, we extract atomic features presented by Pafnucy (Stepniewska-Dziubinska et al., 2018). This atomic feature set was widely used in various structure-based machine learning methods (Jones et al., 2021; Kyro et al., 2023; Shim et al., 2022; 2024). Each atom feature consists of 3D coordinates (x,y,z) of the atom position and its features comprising a 19 dimension feature vector, which includes element type, atom hybridization, number of heavy atom bonds, bond properties (hydrophobic, aromatic, acceptor, donor, ring), partial charge, and Van der Waals radii. All atomic coordinates within each protein-ligand structure are normalized, aligning them with the center of the ligand. We used Open Babel (O’Boyle et al., 2011) for format conversion and partial charge calculation, and utilized the tfbio package (tfbio, 2018) to extract protein, ligand, and binding co-complex features.
2.3 Model architecture
We utilize graph neural network architectures capable of leveraging graph representations of protein-ligand structures to capture both atomic features and their interconnections, which serves as a backbone model. The backbone networks provide low-dimensional
For the backbone graph neural networks, we use atomic-level graph representations, where nodes represent protein or ligand atoms, and edges are connectivity (bonds) between these atoms. Each node encapsulates its corresponding atomic features described earlier. We incorporate both covalent and noncovalent bonds with the Euclidean distances between the atoms. For this graph neural network, we use two network architectures: 1) a variant of spatial-graph convolutional neural network (SGCNN) used in the potential net (Feinberg et al., 2018) and the fusion methods (Jones et al., 2021), and 2) a type of equivariant graph network architectures, E(n) equivariant graph neural networks (EGNN) (Satorras et al., 2021). Note that although we incorporate SGCNN and EGNN into our MIL framework, any graph neural network architectures can be utilized as a backbone model. We utilize multiple graph neural networks, each of which interprets a single instance containing a docking pose with the protein receptor. We incorporate multiple graph networks into an attention network to capture the docking poses close to the target label.
Given a bag of multiple docking pose instances in a low-dimensional embedding, we propose to use attention-network pooling to predict its single binding affinity. Although docking pose instances are generally ordered by docking scores (with the lowest energy pose listed first), the “best” docking poses (closest to the crystal structures or with the highest correlation to experimental binding affinities) are not necessarily the first ones in the set, according to the pose classification studies (Shim et al., 2022; 2024). To address this and to enable permutation-invariant MIL pooling, we use an attention-based weighted average pooling approach, similar to Luong et al. (2015), Ilse et al. (2018). In this model, the weights for each pose instance are determined by the proposed attention network, allowing for order-independent predictions.
The attention network consists of an attention layer, followed by three fully-connected layers. The attention layer serves as the MIL pooling mechanism:
3 Experiment setup
In this section, we describe the experimental setup to demonstrate the effectiveness of our proposed approach. We first detail the datasets and the process for generating molecular docking poses. We provide comprehensive details on the training and evaluation procedures, together with the metrics for quantitative assessment.
3.1 Data
The PDBbind database (Su et al., 2019) is the most widely used dataset in structure-based machine learning studies. It combines experimental binding affinity data (
The PDBbind database provides co-crystal complex structures. Thus, we generate molecular docking poses for each of these crystal structures to demonstrate our MIL-based approach. We employed AutoDock VINA (Trott and Olson, 2010; Eberhardt et al., 2021) to generate docking poses by utilizing the ligand and receptor data with the corresponding binding site. Particularly, we used ConveyorLC (Zhang et al., 2014), a VINA docking pipeline for high-performance computing (HPC) clusters. We generated up to 10 poses for each docking ligand. We also computed the root mean square deviation (RMSD) of atomic positions between each docking pose and its corresponding crystal ligand structure using Open Drug Discovery Toolkit (ODDT) (Wójcikowski et al., 2015). Note that AutoDock VINA computes two RMSD values (lower and upper bounds) relative to the first pose, which is the top-ranked pose with the lowest score, rather than using the crystal structure as a reference. Therefore, we used the ODDT software to independently calculate RMSD values based on the crystal structures. The RMSD values serve as metrics for assessing prediction performance across various structural conformations and poses. Figure 2 shows several examples of docking poses and their RMSDs. In docking poses for the PDBbind dataset, there are cases where all poses are close to the crystal structure (poses with RMSDs below 2 Å), as shown in the top example (“1sqa”) of Figure 2. In contrast, sometimes all poses are inaccurate (poses with RMSDs exceeding 4 Å) (“1pxn”). Another case is that the majority of poses are inaccurate whereas a few pose instances are accurate (“3g0w”). Figure 3 shows the overall RMSD errors of the docking poses in the PDBbind dataset.
Figure 2. Examples of crystal structures and their docking poses of PDBbind CASF-2016 (core set). Autodock VINA (Trott and Olson, 2010; Eberhardt et al., 2021) with ConveyorLC (Zhang et al., 2014) was used for molecular docking.
Figure 3. Statistics of the docking poses for training and evaluation. The plots represent the RMSD histograms between the docking poses and original crystal ligands in PDBbind 2020 general (left), refined (middle), and CASF-2016 (right) sets. Each histogram bin represents the number of poses within the specific RMSD range.
To evaluate our method on cases where co-crystal structures were not available, we applied our method to an additional compound dataset with molecular docking poses. We selected compound collections targeting the SARS-CoV-2 main protease receptor (“Mpro”) curated by Shim et al. (2024), aimed at identifying potential antiviral agents against the virus. The ligand SMILES data with experimental binding affinities were originally sourced from POSTERA (Morris et al., 2021) and GOSTAR (GoStar, 2023). The receptor structure was obtained from the protein data bank (pdbid: 6LU7) (Burley et al., 2019). This dataset does not provide complex crystal structures. Instead, molecular docking poses were generated using ConveyorLC (Zhang et al., 2014) for AutoDock VINA (Trott and Olson, 2010; Eberhardt et al., 2021). For each compound, we generated up to 10 docking poses. For more detail about the docking process of the Mpro dataset, refer to the original paper (Shim et al., 2024).
We used two different splitting methods to divide the entire Mpro compound set into training and test data: random split and scaffold-based split. The scaffold split ensures that similar compound structures are not present in both the training and test sets by grouping the same Bemis-Murckold scaffolds into either the training or testing set. We utilized the scaffold-based method in the ATOM Modeling PipeLine (AMPL) (Minnich et al., 2020) to generate the training and test split. The number of training and test compounds is 2,135, 1,281, respectively.
3.2 Training and evaluation details
We trained and evaluated our proposed models using PyTorch and PyTorch Geometric. Leveraging their distributed data parallel features, we utilized a cluster of 4 NVIDIA Titan Xp GPUs, each of which has 12 GB of memory. There are two options to train our MIL models: 1) simultaneous training: train both the backbone graph neural network and the MIL attention network together in a unified process, 2) separate training: first, train the backbone models independently and freeze the model parameters to extract features (low-dimensional embedding for 3D complexed pose structures). These features are then used to train the MIL attention network separately. For simultaneous training, we used a mini-batch size of 1 due to GPU memory constraints. While this approach can offer slightly improved predictions for some data, it is computationally expensive because it processes multiple docking pose instances (around 10 poses) with the backbone graph networks simultaneously. In contrast, the separate training option is more computationally efficient. By training the MIL attention network independently, we significantly reduced GPU resource usage and achieved more efficient computation. We used option 2 (separate training) for our experiments.
We used the RMSprop optimizer with the multi-step learning rate scheduler. The total number of epochs is 100-200, depending on the convergence of the loss function. The mini-batch size is 20. Starting with an initial learning rate of 0.001, it decreases by a factor of 0.9 at epoch intervals of 5, 10, 20, 30, 50, 70, and 90. We chose not to utilize checkpoint selection based on the validation split. Instead, we simply opted for the latest checkpoint available for our evaluation.
The hyperparameters for the SGCNN model are as follows: the thresholds for covalent and non-covalent neighbors are set to 1.5 and 4.5 Å, respectively. The size of the gated graph sequence for both covalent and non-covalent bonds is 2. The “gather” widths for covalent and non-covalent bonds are 16 and 12, respectively. We used the following hyperparameters for the EGNN model: the number of edge features is 1, and the model has 6 layers. The distance cut-off is set to 5 Å. Residual connections are used together with attention layers.
4 Results
We performed experiments to demonstrate the effectiveness of the proposed MIL approach using the molecular docking pose structures. Our evaluation report on the performance of our proposed approach with other existing methods used two datasets with their Autodock VINA docking poses. We present the following quantitative metrics: root mean square error (RMSE), mean absolute error (MAE), coefficient of determination
We benchmarked our results with those from other recent 3D structure-based methods: 1) SGCNN, a variant of Potential Net (Feinberg et al., 2018; Jones et al., 2021), 2) EGNN, equivariant graph neural network, and 3) HACNET (Kyro et al., 2023), which incorporates hybrid attentions through 3DCNN and GCN. We chose SGCNN and EGNN for comparative analysis because our method integrates the architecture into the MIL framework as backbone models. This choice allows us to analyze the behavior of our proposed MIL mechanism. Additionally, we utilized the model checkpoint for HACNET trained on crystal structures in the PDBbind 2016 general and refined sets, due to unavailability of model re-training using the PDBbind 2020 dataset. These comparative models predict binding affinities based on individual docking pose instances. One could report the quantitative metrics using all individual pose results for the comparative models. However, this method does not align with our MIL approach, which generates a single binding affinity from multiple docking poses for each complex entity. Instead, we use the following two methods. First, we report the metrics based on the predicted binding affinities for the top poses, specifically the first pose with the lowest docking scores (“Top”). This approach is commonly used in high-throughput screening with molecular docking tools. Second, we group the results based on compound IDs (e.g., PDB IDs in the PDBbind dataset), and use the average of predicted binding affinities for the same complex entity to ensure consistency with our results (“Avg”).
4.1 PDBbind
Table 1 summarizes the model performance on docking poses with the crystal structures in the PDBbind CASF-2016 (core set). Each complex entity can have up to 11 pose instances, including the crystal ligand instance with up to 10 docking poses. The first group (row 1–6) in Table 1 presents results from models trained exclusively on crystal structures. The SGCNN and EGNN models (row 3–6) were trained using crystal structures from the PDBbind 2020 general and refined sets. In contrast, we utilized the HACNET model (row 1–2) trained on crystal structures from the PDBbind 2016 general and refined sets. The second and third groups (row 7-9 and 10-12, respectively) show results from models trained on both crystal structures and docking poses from the PDBbind 2020 general and refined sets. “Top” presents results based on the top poses, while “Avg” shows results based on non-weighted averages across multiple docking poses within the same complex entities. The SGCNN-MIL and EGNN-MIL results present our proposed MIL method integrating the spatial graph network (SGCNN) and the equivariant graph network (EGNN) as backbone models, respectively.
Table 1. Model Performance of Binding Affinity Prediction on docking and crystal structures of the PDBbind CASF-2016. The models in the first group (row 1–6) were trained solely on the crystal structures of the PDBbind general and refined sets. HACNET was trained using the PDBbind 2016 dataset, while the other models were trained on the PDBbind 2020 dataset. The models in the second and third groups (row 7-9, and row 10–12) were trained on both docking and crystal structures of the PDBbind general and refined sets.
We conducted an additional experiment to show the model performance based training solely on the docking poses without the crystal ligand structures from the PDBbind CASF-2016 (core set). Table 2 shows the model performance among several methods. Similar to the results in Table 1, the first 7 rows present results from HACNET, SGCNN, EGNN, and late-fusion models, respectively, trained exclusively on crystal structures. The HACNET and fusion models (row 1,2 and 7) were trained on crystal structures from the PDBbind 2016 general and refined sets. In contrast, the SGCNN and EGNN models (row 3,4,5 and 6) were trained on crystal structures from the PDBbind 2020 general and refined sets. The second and third groups (row 8-10 and 11-13, respectively) show results from models trained exclusively on docking poses from the PDBbind 2020 general and refined sets. For each of the two backbone models (SGCNN and EGNN), we report the top-pose only (“Top”), averaged (“Avg”) across the poses, and our (“MIL”) results.
Table 2. Model Performance of Binding Affinity Prediction exclusively on docking structures of the PDBbind CASF-2016. The training sets used for the first group follow the same setup as in the previous experiment (Table 1). However, models in the second and third groups were trained solely on docking poses, without crystal structures.
Results in Tables 1, 2 show that our proposed method (“MIL”) outperforms the other methods in both cases. Overall, models trained on both crystal structures and docking poses, or solely on docking poses, outperform those trained exclusively on crystal structures. Among the SGCNN-based models (second group), our MIL method significantly outperforms the other approaches. The results from the third group (EGNN) also demonstrate that our method outperforms the other approaches.
4.2 SARS-CoV-2 Mpro
We conducted experiments using the SARS-CoV-2 (COVID-19) main protease dataset (“Mpro”), sourced from POSTERA (Morris et al., 2021) and GOSTAR (GoStar, 2023), and curated and docking processed by Shim et al. (2024). Our evaluations include both random splits and scaffold splits of the “Mpro” dataset. Tables 3, 4 summarize the model performance in predicting binding affinities on the dataset with random and scaffold splits, respectively.
Table 3. Model Performance of Binding Affinity Prediction on SARS-CoV-2 “Mpro” dataset with a random split.
Table 4. Model Performance of Binding Affinity Prediction on SARS-CoV-2 “Mpro” dataset with a scaffold split.
The evaluation results with the random split, as shown in Table 3, indicate that our proposed method generally outperforms the approaches using only the top docking pose (“Top”) or the averaged pose (“Avg”) results. However, it does not achieve the same level of performance as observed in the PDBbind evaluations. In the scaffold split scenario, where no similar compounds overlap between the training and evaluation sets, our approach clearly outperforms the other methods.
5 Ablation study and discussion
Our experiments with the PDBbind and Mpro datasets demonstrate that our proposed MIL approach consistently outperforms existing methods. We discuss several key observations from these experiments. First, for each backbone architecture, our MIL approach produces more accurate predictions than the non-weighted averages of the predicted binding affinities from multiple docking pose instances and the predictions from the top-ranked poses. The results demonstrate that the attention network module of our approach effectively predicts accurate binding affinities across multiple pose instances within each complex entity. In comparing the EGNN-MIL models trained on both the crystal and docking pose data with those trained on docking pose data alone, there is no significant discrepancy, as illustrated in Tables 1, 2). The model trained on both crystal and docking data exhibits slightly improved RMSE, MAE, and
Second, incorporating docking poses generally improve the model prediction over training models solely on crystal structure data. For instance, SGCNN-Top and SGCNN-Avg (rows 7-8 vs. row 3–4) and EGNN-Top and EGNN-Avg (rows 10-11 vs. row 5–6) in Table 2). demonstrate this clearly. It can be argued that incorporating docking data introduces greater diversity into the data distribution, enhancing the model’s predictive performance and generalizability. In addition, adding crystal structures into the training data yields a slight improvement in model accuracy, as shown in a comparison of the results in Table 1 with those in Table 2. However, results may vary based on the quality of the docking poses for each complex structure entity. If all docking poses are significantly inaccurate relative to the crystal structures, incorporating crystal structures could enhance model accuracy.
Furthermore, the choice of backbone graph network architectures also affects the prediction accuracy. The EGNN models demonstrate a clear advantage over the SGCNN models, likely due to its effective feature learning capabilities, which enhance understanding of the input geometry of ligand-protein structures. This results in improved generalizability across different configurations (poses) within the same underlying co-complex entity.
The results evaluating the SARS-CoV-2 Mpro dataset indicate that our MIL approach achieves improved accuracy, compared to the other methods. However, the difference between the correlation coefficients (Pearson and Spearman) becomes marginal in the random split scenario. In contrast, our approach clearly outperforms other methods in the scaffold split scenario. In comparison to the results for the PDBbind dataset, the models trained on the Mpro dataset yield relatively poor performance. It is possibly due to greater inaccuracies in the Mpro docking poses compared to those in the PDBbind dataset.
5.1 Performance with good poses
We evaluate model performance based on the number of good docking poses within each complex entity. We define “good” docking poses as those with an RMSD less than 2.5 Å. We categorize the evaluation data into three groups: structures with no good poses, those with 1–4 good poses, and those with more than 4 good poses. Figure 4 shows model performance based on the number of good poses in the PDBbind CASF-2016 set. Compared to models that rely on top poses (Top) or average predictions across multiple poses (Avg), our approach consistently demonstrates improved and consistent performance. Notably, groups with a sufficient number of good poses tend to outperform those lacking good poses.
Figure 4. Model performance of binding affinity prediction on the PDBbind CASF-2016 dataset, based on the number of good docking poses within each complex entity. The evaluation data were divided into three groups: structures with no good poses, those with 1–4 good poses, and those with more than 4 good poses. The top and bottom row show results using the SGCNN and EGNN backbones, respectively. The left and right columns show RMSE and
As we discussed earlier, models trained on the Mpro dataset generally underperform, compared to those trained on the PDBbind dataset. One possible explanation is that the quality of the docking poses (the number of “good” poses for each complex entity) may be lower in the Mpro dataset than in the PDBbind dataset. However, analyzing the quality of the docking poses in the Mpro dataset is non-trivial, due to the absence of co-crystal structures. Instead, we measure the RMSD errors between pairs of docking poses within each complex entity. Figure 5 shows the histograms of RMSD errors for pairwise ligand poses in both datasets. While these histograms do not reflect accurate poses based on crystal structures, they indicate the consistency of the docking poses for each complex. The docking poses in the Mpro dataset noticeably show greater discrepancies compared to those in the PDBbind CASF-2016 set. In real compound screening, compounds with highly inconsistent docking poses may be excluded from further ML predictions. Moreover, utilizing more advanced, state-of-the-art molecular docking simulations could improve overall prediction accuracy together with our MIL approach.
Figure 5. Histograms (probability density) of pair-wise RMSD errors between ligand docking poses in the PDBbind CASF-2016 (left) and SARS-CoV-2 Main Protease (right) datasets.
5.2 Optimal number of docking poses
In multi-instance learning, finding an optimal range of the bag size
Figure 6. Model performance of binding affinity prediction on the PDBbind CASF-2016 dataset, based on the number of docking pose instances within each complex entity. The evaluation was performed where the number of poses is 1, 3, 5, 7, and 9. The top and bottom row show results using the SGCNN and EGNN backbones, respectively. The left and right columns show RMSE and
5.3 Attention types
We evaluated two different attention types for multi-instance learning. The first method involves global attention for weighted average pooling, as detailed in Section 2. The second method utilizes a multi-head attention mechanism (Vaswani et al., 2017). We observed no significant change or improvement using the multi-head attention network. While the multi-head attention mechanism offers slight improvements in predictions for SGCNN-MIL models (RMSE, MAE and
Table 5. Model Performance in Binding Affinity Prediction on PDBbind CASF-2016 using two types of attention mechanisms.
5.4 Computational Costs
The attention network in our MIL framework is lightweight, comprising approximately 68,000 parameters, which is under 1 MB in size. The training and evaluation of our models are fast. Using a single NVIDIA Titan Xp GPU, the training time for one epoch with a batch size of 20, across 18,539 complex entities with up to 10 docking poses, is approximately 36.5 s (1.97 milliseconds per complex entity).
6 Conclusion
We presented a new structure-based multi-instance learning approach utilizing molecular docking poses as a bag of complex entity. Our method employs an attention network together with graph neural networks to enable permutation-invariant weighted average pooling between docking poses. We demonstrated our method using the PDBbind and the Mpro datasets. Our method offers binding affinity prediction without requiring co-crystal structures, which increases its applicability for various targets of interest.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.
Author contributions
HK: Conceptualization, Data curation, Formal Analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing–original draft, Writing–review and editing. HS: Data curation, Investigation, Writing–review and editing, Software. AR: Software, Writing–review and editing. SH: Resources, Software, Writing–review and editing. GS: Resources, Writing–review and editing. JA: Funding acquisition, Project administration, Supervision, Writing–review and editing.
Funding
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was funded by the Defense Threat Reduction Agency (DTRA), HDTRA1242044 (HK, SH, GS) and HDTRA1036045 (HS, AR, JA).
Acknowledgments
This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under contract DE-AC52-07NA27344. Lawrence Livermore National Security, LLC. LLNL-JRNL-2000673.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Backenköhler, M., Groß, J., Wolf, V., and Volkamer, A. (2024). Guided docking as a data generation approach facilitates structure-based machine learning on kinases. J. Chem. Inf. Model. 64, 4009–4020. doi:10.1021/acs.jcim.4c00055
Burley, S. K., Berman, H. M., Bhikadiya, C., Bi, C., Chen, L., Costanzo, L. D., et al. (2019). Protein data bank: the single global archive for 3d macromolecular structure data. Nucleic Acids Res. 47, D520–D528. doi:10.1093/nar/gky949
Carbonneau, M.-A., Cheplygina, V., Granger, E., and Gagnon, G. (2018). Multiple instance learning: a survey of problem characteristics and applications. Pattern Recognit. 77, 329–353. doi:10.1016/j.patcog.2017.10.009
Chithrananda, S., Grand, G., and Ramsundar, B. (2020). Chemberta: large-scale self-supervised pretraining for molecular property prediction. Corr. abs/2010, 09885. doi:10.48550/arXiv.2010.09885
Corso, G., Stärk, H., Jing, B., Barzilay, R., and Jaakkola, T. (2023). “Diffdock: diffusion steps, twists, and turns for molecular docking,” in International conference on learning representations (ICLR).
Dooly, D. R., Zhang, Q., Goldman, S. A., and Amar, R. A. (2002). Multiple instance learning of real valued data. J. Mach. Learn. Res. 3, 651–678.
Duvenaud, D., Maclaurin, D., Aguilera-Iparraguirre, J., Gómez-Bombarelli, R., Hirzel, T., Aspuru-Guzik, A., et al. (2015). “Convolutional networks on graphs for learning molecular fingerprints,” in Proceedings of the 28th international conference on neural information processing systems - volume 2 (Cambridge, MA, USA: MIT Press), NIPS’15), 2224–2232.
Eberhardt, J., Santos-Martins, D., Tillack, A. F., and Forli, S. (2021). Autodock vina 1.2.0: new docking methods, expanded force field, and python bindings. J. Chem. Inf. Model. 61, 3891–3898. doi:10.1021/acs.jcim.1c00203
Fatima, S., Ali, S., and Kim, H.-C. (2023). A comprehensive review on multiple instance learning. Electronics 12, 4323. doi:10.3390/electronics12204323
Feinberg, E. N., Sheridan, R., Joshi, E., Pande, V. S., and Cheng, A. C. (2019). Step change improvement in admet prediction with potentialnet deep featurization. arXiv:1903. doi:10.48550/arXiv.1903.11789
Feinberg, E. N., Sur, D., Wu, Z., Husic, B. E., Mai, H., Li, Y., et al. (2018). Potentialnet for molecular property prediction. ACS Cent. Sci. 4, 1520–1530. doi:10.1021/acscentsci.8b00507
Friesner, R. A., Banks, J. L., Murphy, R. B., Halgren, T. A., Klicic, J. J., Mainz, D. T., et al. (2004). Glide: a new approach for rapid, accurate docking and scoring. 1. method and assessment of docking accuracy. J. Med. Chem. 47, 1739–1749. doi:10.1021/jm0306430
Fuchs, F. B., Worrall, D. E., Fischer, V., and Welling, M. (2020). “Se(3)-transformers: 3d roto-translation equivariant attention networks,” in Proceedings of the 34th international conference on neural information processing systems (NIPS’20: Curran Associates Inc).
Gomes, J., Ramsundar, B., Feinberg, E., and Pande, V. (2017). Atomic convolutional networks for predicting protein-ligand binding affinity.
Halgren, T. A., Murphy, R. B., Friesner, R. A., Beard, H. S., Frye, L. L., Pollard, W. T., et al. (2004). Glide: a new approach for rapid, accurate docking and scoring. 2. enrichment factors in database screening. J. Med. Chem. 47, 1750–1759. doi:10.1021/jm030644s
Ilse, M., Tomczak, J. M., and Welling, M. (2018). Attention-based deep multiple instance learning. arXiv Prepr. arXiv:1802.04712. doi:10.48550/arXiv.1802.04712
Jiang, D., Hsieh, C.-Y., Wu, Z., Kang, Y., Wang, J., Wang, E., et al. (2021). Interactiongraphnet: a novel and efficient deep graph representation learning framework for accurate protein–ligand interaction predictions. J. Med. Chem. 64, 18209–18232. doi:10.1021/acs.jmedchem.1c01830
Jiménez, J., Doerr, S., Martínez-Rosell, G., Rose, A. S., and De Fabritiis, G. (2017). Deepsite: protein-binding site predictor using 3d-convolutional neural networks. Bioinformatics 33, 3036–3042. doi:10.1093/bioinformatics/btx350
Jiménez, J., Škalič, M., Martínez-Rosell, G., and De Fabritiis, G. (2018). Kdeep: protein-ligand absolute binding affinity prediction via 3d-convolutional neural networks. J. Chem. Inf. Model. 58, 287–296. doi:10.1021/acs.jcim.7b00650
Jones, D., Kim, H., Zhang, X., Zemla, A., Stevenson, G., Bennett, W. F. D., et al. (2021). Improved protein–ligand binding affinity prediction with structure-based deep fusion inference. J. Chem. Inf. Model. 61, 1583–1592. doi:10.1021/acs.jcim.0c01306
Kyro, G. W., Brent, R. I., and Batista, V. S. (2023). Hac-net: a hybrid attention-based convolutional neural network for highly accurate protein-ligand binding affinity prediction. J. Chem. Inf. Model. 63, 1947–1960. doi:10.1021/acs.jcim.3c00251
Lu, W., Wu, Q., Zhang, J., Rao, J., Li, C., and Zheng, S. (2022). “Tankbind: trigonometry-aware neural networks for drug-protein binding structure prediction,” in Proceedings of the 36th international conference on neural information processing systems.
Luong, T., Pham, H., and Manning, C. D. (2015). “Effective approaches to attention-based neural machine translation,” in Proceedings of the 2015 conference on empirical methods in natural language processing, 1412–1421. doi:10.18653/v1/D15-1166
Mazuz, E., Shtar, G., Kutsky, N., Rokach, L., and Shapira, B. (2023). Pretrained transformer models for predicting the withdrawal of drugs from the market. Bioinformatics 39, btad519. doi:10.1093/bioinformatics/btad519
McNutt, A., Francoeur, P., Aggarwal, R., Masuda, T., Meli, R., Ragoza, M., et al. (2021). Gnina 1.0: molecular docking with deep learning. J. Cheminformatics 13, 43. doi:10.1186/s13321-021-00522-2
Méndez-Lucio, O., Ahmad, M., Rio-Chanona, E. A. d., and Wegner, J. K. (2021). A geometric deep learning approach to predict binding conformations of bioactive molecules. Nat. Mach. Intell. 3, 1033—–1039. doi:10.1038/s42256-021-00409-9
Minnich, A. J., McLoughlin, K., Tse, M., Deng, J., Weber, A., Murad, N., et al. (2020). Ampl: a data-driven modeling pipeline for drug discovery. J. Chem. Inf. Model. 60, 1955–1968. doi:10.1021/acs.jcim.9b01053
Moriwaki, H., Tian, Y.-S., Kawashita, N., and Takagi, T. (2018). Mordred: a molecular descriptor calculator. J. Cheminf 10, 4. doi:10.1186/s13321-018-0258-y
Morris, A., McCorkindale, W., Consortium, T., Drayman, N., Chodera, J., Tay, S., et al. (2021). Discovery of sars-cov-2 main protease inhibitors using a synthesis-directed de novo design model. Chem. Commun. 57, 5909–5912. doi:10.1039/d1cc00050k
Mqawass, G., and Popov, P. (2024). graphlambda: fusion graph neural networks for binding affinity prediction. J. Chem. Inf. Model. 64, 2323–2330. doi:10.1021/acs.jcim.3c00771
O’Boyle, N. M., Banck, M., James, C. A., Morley, C., Vandermeersch, T., and Hutchison, G. R. (2011). Open babel: an open chemical toolbox. J. Cheminf. 3, 33. doi:10.1186/1758-2946-3-33
Powers, A. S., Yu, H. H., Suriana, P., and Dror, R. O. (2022). Fragment-based ligand generation guided by geometric deep learning on protein-ligand structure. bioRxiv. doi:10.1101/2022.03.17.484653
Ragoza, M., Hochuli, J., Idrobo, E., Sunseri, J., and Koes, D. R. (2017). Protein–ligand scoring with convolutional neural networks. J. Chem. Inf. Model. 57, 942–957. doi:10.1021/acs.jcim.6b00740
Ramsundar, B., Eastman, P., Walters, P., Pande, V., Leswing, K., and Wu, Z. (2019). Deep Learning for the life Sciences (O’reilly media). Available at: https://www.amazon.com/Deep-Learning-Life-Sciences-Microscopy/dp/1492039837.
Rogers, D., and Hahn, M. (2010). Extended-connectivity fingerprints. J. Chem. Inf. Model. 50, 742–754. doi:10.1021/ci100050t
Rong, Y., Bian, Y., Xu, T., Xie, W., Wei, Y., Huang, W., et al. (2020). Self-supervised graph transformer on large-scale molecular data. Adv. Neural Inf. Process. Syst. 33. doi:10.48550/arXiv.2007.02835
Satorras, V. G., Hoogeboom, E., and Welling, M. (2021). “E(n) equivariant graph neural networks,” in Proceedings of the 38th international conference on machine learning. (PMLR), vol. 139 of Proceedings of machine learning research. Editors M. Meila, and T. Zhang 9323–9332.
Scantlebury, J., Vost, L., Carbery, A., Hadfield, T. E., Turnbull, O. M., Brown, N., et al. (2023). A small step toward generalizability: training a machine learning scoring function for structure-based virtual screening. J. Chem. Inf. Model. 63, 2960–2974. doi:10.1021/acs.jcim.3c00322
Shim, H., Allen, J. E., and Bennett, W. F. D. (2024). Enhancing docking accuracy with pecan2, a 3d atomic neural network trained without co-complex crystal structures. Mach. Learn. Knowl. Extr. 6, 642–657. doi:10.3390/make6010030
Shim, H., Kim, H., Allen, J. E., and Wulff, H. (2022). Pose classification using three-dimensional atomic structure-based neural networks applied to ion channel–ligand docking. J. Chem. Inf. Model. 62, 2301–2315. doi:10.1021/acs.jcim.1c01510
Stärk, H., Beaini, D., Corso, G., Tossou, P., Dallago, C., Günnemann, S., et al. (2022a). “3d infomax improves gnns for molecular property prediction,” in Proceedings of the 39th international Conference on machine learning. Vol. 162 of Proceedings of machine learning research, 20479–20502.
Stärk, H., Ganea, O., Pattanaik, L., Barzilay, R., and Jaakkola, T. (2022b). “Equibind: geometric deep learning for drug binding structure prediction,” in International Conference on machine learning (PMLR), 20503–20521.
Stepniewska-Dziubinska, M. M., Zielenkiewicz, P., and Siedlecki, P. (2018). Development and evaluation of a deep learning model for protein–ligand binding affinity prediction. Bioinformatics 34, 3666–3674. doi:10.1093/bioinformatics/bty374
Su, M., Yang, Q., Du, Y., Feng, G., Liu, Z., Li, Y., et al. (2019). Comparative assessment of scoring functions: the casf-2016 update. J. Chem. Inf. Model. 59, 895–913. doi:10.1021/acs.jcim.8b00545
Trott, O., and Olson, A. J. (2010). Autodock vina: improving the speed and accuracy of docking with a new scoring function, efficient optimization and multithreading. J. Comput. Chem. 31, 455–461. doi:10.1002/jcc.21334
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2017). “Attention is all you need,”. Advances in neural information processing systems. Editors I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathanet al. (NY, United States: Curran Associates, Inc.), 30.
Wójcikowski, M., Zielenkiewicz, P., and Siedlecki, P. (2015). Open drug discovery toolkit (oddt): a new open-source player in the drug discovery field. J. Cheminformatics 7. 26. doi:10.1186/s13321-015-0078-2
Zhang, H., Liao, L., Saravanan, K. M., Yin, P., and Wei, Y. (2019). Deepbindrg: a deep learning based method for estimating effective protein-ligand affinity. PeerJ 7, e7362. doi:10.7717/peerj.7362
Zhang, S., Jin, Y., Liu, T., Wang, Q., Zhang, Z., Zhao, S., et al. (2023). Ss-gnn: a simple-structured graph neural network for affinity prediction. ACS Omega 8, 22496–22507. doi:10.1021/acsomega.3c00085
Zhang, X., Wong, S. E., and Lightstone, F. C. (2014). Toward fully automated high performance computing drug discovery: a massively parallel virtual screening pipeline for docking and molecular mechanics/generalized born surface area rescoring to improve enrichment. J. Chem. Inf. Model. 54, 324–337. doi:10.1021/ci4005145
Keywords: AI-driven drug development, virtual high-throughput screening, protein-ligand interaction, molecular docking, 3D atomic graph representation, structure-based machine learning, multi-instance learning, attention mechanism
Citation: Kim H, Shim H, Ranganath A, He S, Stevenson G and Allen JE (2025) Protein-ligand binding affinity prediction using multi-instance learning with docking structures. Front. Pharmacol. 15:1518875. doi: 10.3389/fphar.2024.1518875
Received: 29 October 2024; Accepted: 26 November 2024;
Published: 03 January 2025.
Edited by:
Moom Rahman Roosan, Chapman University, United StatesCopyright © 2025 Kim, Shim, Ranganath, He, Stevenson and Allen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Hyojin Kim, aGtpbUBsbG5sLmdvdg==