Skip to main content

ORIGINAL RESEARCH article

Front. Pharmacol., 03 January 2025
Sec. Experimental Pharmacology and Drug Discovery
This article is part of the Research Topic Machine Learning Advancements in Pharmacology: Transforming Drug Discovery and Healthcare View all 5 articles

Protein-ligand binding affinity prediction using multi-instance learning with docking structures

Hyojin Kim
Hyojin Kim1*Heesung ShimHeesung Shim2Aditya RanganathAditya Ranganath1Stewart HeStewart He3Garrett StevensonGarrett Stevenson4Jonathan E. AllenJonathan E. Allen3
  • 1Center for Applied Scientific Computing, Lawrence Livermore National Laboratory, Livermore, CA, United States
  • 2Biosciences and Biotechnology Division, Lawrence Livermore National Laboratory, Livermore, CA, United States
  • 3Global Security Computing Applications Division, Lawrence Livermore National Laboratory, Livermore, CA, United States
  • 4Computational Engineering Division, Lawrence Livermore National Laboratory, Livermore, CA, United States

Introduction: Recent advances in 3D structure-based deep learning approaches demonstrate improved accuracy in predicting protein-ligand binding affinity in drug discovery. These methods complement physics-based computational modeling such as molecular docking for virtual high-throughput screening. Despite recent advances and improved predictive performance, most methods in this category primarily rely on utilizing co-crystal complex structures and experimentally measured binding affinities as both input and output data for model training. Nevertheless, co-crystal complex structures are not readily available and the inaccurate predicted structures from molecular docking can degrade the accuracy of the machine learning methods.

Methods: We introduce a novel structure-based inference method utilizing multiple molecular docking poses for each complex entity. Our proposed method employs multi-instance learning with an attention network to predict binding affinity from a collection of docking poses.

Results: We validate our method using multiple datasets, including PDBbind and compounds targeting the main protease of SARS-CoV-2. The results demonstrate that our method leveraging docking poses is competitive with other state-of-the-art inference models that depend on co-crystal structures.

Discussion: This method offers binding affinity prediction without requiring co-crystal structures, thereby increasing its applicability to protein targets lacking such data.

1 Introduction

Given the vast number of drug-like chemical compounds and the rapid growth of compound library databases, it is crucial to develop an effective tool for large-scale virtual high-throughput screening to identify hit molecules. To address this need, recent studies have employed various machine learning and deep learning-based inference models to enhance traditional computational modeling techniques such as molecular docking and molecular dynamics simulations for accurately predicting protein-ligand interactions and binding affinities.

One major application utilizing machine learning techniques in drug discovery involves the use of SMILES strings and their associated features such as molecular descriptors (RDKit, 2024; Moriwaki et al., 2018) and fingerprints (Rogers and Hahn, 2010), to predict compound properties and target-specific interactions. This includes predicting toxicity, solubility, membrane permeability, binding affinity, and other compound characteristics associated with the target receptors. The approaches in this category use a variety of machine learning methods, ranging from random forest algorithms to various neural network architectures (Minnich et al., 2020; Ramsundar et al., 2019).

Recent studies use advanced architectures and techniques with atomic graph representations such as graph neural networks (Duvenaud et al., 2015; Stärk et al., 2022a), graph transformers and attention mechanisms (Rong et al., 2020) in unsupervised or self-supervised settings. These approaches show the advantages of utilizing pretrained models on unlabeled large-volume compound databases. Furthermore, large language model-based approaches have been proposed to address pretrainng and finetuning with the unlabeled compound data (Chithrananda et al., 2020; Mazuz et al., 2023).

Another direction in leveraging deep learning techniques uses 3D atomic representations of the protein-ligand structures to predict the interaction and binding affinities. These structure-based methods use 3D convolutional neural networks (3DCNNs) utilizing voxelized atomic representations of co-crystal complex structures to predict binding affinities (Gomes et al., 2017; Ragoza et al., 2017; Jiménez et al., 2018; Zhang et al., 2019) and to predict protein-binding sites (Jiménez et al., 2017). Furthermore, graph neural networks (GNNs), graph convolutional networks (GCNs), and spatial graph convolutional neural networks (SGCNNs) have been introduced for binding affinity prediction. These models utilize graph representations of atomic data from co-crystal complex structures, capturing atoms and their connectivity through nodes and edges (Feinberg et al., 2018; 2019; Zhang et al., 2023). Recently, equivariant attention networks have received attention for representing 3D point cloud or graph data under rotations such as SE(3)-transformers (Fuchs et al., 2020). Satorras et al. developed light-weight E(n) equivariant graph neural networks (EGNNs) equivariant to rotations, translations, reflections and permutations without the need for higher-order representations (Satorras et al., 2021). Powers et al. proposed to use data-efficient E(3) equivariant networks with point cloud data to guide how to attach new molecular fragments in molecule optimization (Powers et al., 2022). Scantlebury et al. proposed a lightweight E(n)-equivariant graph neural network for machine learning-based scoring functions (Scantlebury et al., 2023).

An additional recent trend involves using hybrid neural network architectures or fusion networks to capture diverse input representations and interactions. Jones et al. proposed a fusion method to complement 3DCNN and SGCNN architectures to effectively capture both spatial information and atom connectivities (Jones et al., 2021). Jiang et al. proposed the use of two independent graph network modules to capture intra- and intermolecular interactions (Jiang et al., 2021). Kyro et al. proposed the use of hybrid attentions with 3DCNN and GCN (Kyro et al., 2023). Mqawass et al. proposed a fusion method of different input representations (Mqawass and Popov, 2024).

Most methods in this category rely on co-crystal complex structures of protein-ligand data to train the neural network models. However, crystal structures are not readily available in the early stages of most drug discovery applications. The PDBbind dataset (Su et al., 2019), a meticulously curated subset of the Protein Data Bank (PDB) (Burley et al., 2019) containing experimental binding affinity data, is virtually the sole dataset accessible for structure-based protein-ligand learning tasks. For that reason, we often utilize simulated docking poses of protein-ligand complex structures for structure-based deep learning inference models. The docking poses can be generated using molecular docking tools such as AutoDock VINA (Trott and Olson, 2010; Eberhardt et al., 2021), and GLIDE (Friesner et al., 2004; Halgren et al., 2004). Recent studies have proposed the use of deep neural network approaches to serve as docking scoring functions (McNutt et al., 2021), or to generate docking poses including DeepDock (Méndez-Lucio et al., 2021), EquiBind (Stärk et al., 2022b), TankBind (Lu et al., 2022), DiffDock (Corso et al., 2023), and UniMol (Zhou et al., 2023).

While generating docking poses is feasible and practical, training or evaluating deep neural network models on these poses results in inaccurate predictions, as the docking poses are considered “noisy” data. To identify inaccurate docking poses, Shim et al. proposed a pose classification approach using two different neural network methods, 3DCNN and point cloud network (PCN) (Shim et al., 2022). While the method filters out inaccurate docking poses; it requires an additional computational or inference method for binding affinity estimation. Another study developed a pose classification method without using crystal structures to improve the correlation between docking scores and experimental data (Shim et al., 2024). However, their method relies on the docking poses and their scoring functions. A recently proposed method leverages an E(3)-invariant graph neural network, combined with docking and binding affinity data, to address binding affinity prediction for kinase compounds (Backenköhler et al., 2024). Although this approach utilizes template docking to achieve more accurate poses, it does not explicitly address the distinction between accurate and inaccurate docking data, particularly in scenarios where a significant amount of inaccurate docking data is present.

In this paper, we present a novel inference method based on multi-instance learning (MIL) that utilizes a set of docking poses for each protein-ligand entity to predict binding affinity. Multi-instance learning is a type of weakly supervised learning used to predict a label for a group (bag) of instances, where the individual labels of the instances within the bag are not available (Carbonneau et al., 2018). In the standard MIL setup for classification, it is assumed that negative bags exclusively contain negative instances, while positive bags contain at least one positive instance. This technique offers flexibility by mitigating the potential problems associated with ambiguous and time-consuming labeling, compared to the traditional supervised learning. For more detail about MIL methods in various computer vision and machine learning applications, refer to the following review papers (Carbonneau et al., 2018; Fatima et al., 2023).

We leverage MIL-based regression to predict binding affinities when only molecular docking data is available, as there are no accessible co-crystal structures. Our method begins with graph neural networks, specifically SGCNN and EGNN, representing each ligand pose alongside its associated target receptor instance as a graph embedding. We then extract latent features from multiple pose instances per complex entity using these graph networks. Finally, we integrate these features into our pose-wise attention network, which functions as a MIL framework to predict binding affinities for each complex entity. We train and evaluate our models on molecular docking structures that may contain inaccurate poses. We demonstrate our method using the PDBbind and SARS-CoV-2 main protease datasets. The primary contribution of our method is as follows:

Our method does not rely on co-crystal structures, which increases applicability across various drug target applications.

Multiple docking poses per complex structure entity can be utilized together, without the need to filter out inaccurate ones.

Our pose-wise attention mechanism, which can be integrated with various backbone network architectures, enhances binding affinity prediction even with docking pose data only, compared to existing scoring methods including top-pose-based and non-weighted average scores.

2 Methods

2.1 Problem definition

Most high-throughput screening and 3D structure-based neural network approaches utilize co-crystal complex structures or the “best” poses with the highest docking scores (lowest binding energy). Instead, we consider all generated poses with the target receptor as a bag for each complex structure entity. We formulate this problem as multi-instance learning (MIL) regression to predict a single real-value label for a bag (Dooly et al., 2002). Each bag represents a single protein-ligand structure entity consisting of multiple docking poses, X={x1,,xK}, with at least one instance being close to the target label y. It is assumed that the sequence of docking pose instances can be random. The real-value labels represent the binding affinities. Each instance xiRD represents the atomic representation of the docking pose with the protein receptor, including the 3D coordinates of the atoms and their atomic features. The number of docking poses K can vary depending on the molecular docking methods, generally ranging from 10 to 20 in our applications. Our goal is to predict the binding affinity y by leveraging our proposed MIL network framework M together with the backbone graph network G to interpret a set of docking pose and the receptor instances X.

2.2 Input data and featurization

For input ligand docking poses and protein receptors, we extract atomic features presented by Pafnucy (Stepniewska-Dziubinska et al., 2018). This atomic feature set was widely used in various structure-based machine learning methods (Jones et al., 2021; Kyro et al., 2023; Shim et al., 2022; 2024). Each atom feature consists of 3D coordinates (x,y,z) of the atom position and its features comprising a 19 dimension feature vector, which includes element type, atom hybridization, number of heavy atom bonds, bond properties (hydrophobic, aromatic, acceptor, donor, ring), partial charge, and Van der Waals radii. All atomic coordinates within each protein-ligand structure are normalized, aligning them with the center of the ligand. We used Open Babel (O’Boyle et al., 2011) for format conversion and partial charge calculation, and utilized the tfbio package (tfbio, 2018) to extract protein, ligand, and binding co-complex features.

2.3 Model architecture

We utilize graph neural network architectures capable of leveraging graph representations of protein-ligand structures to capture both atomic features and their interconnections, which serves as a backbone model. The backbone networks provide low-dimensional K embedding representation H={h1,,hK} for a bag of multiple docking pose instances. Thus, hk=Gψ(xk), where Gψ() is a graph neural network with parameters ψ. The low-dimensional embedding H is used as input features in the MIL-based attention network for final prediction.

For the backbone graph neural networks, we use atomic-level graph representations, where nodes represent protein or ligand atoms, and edges are connectivity (bonds) between these atoms. Each node encapsulates its corresponding atomic features described earlier. We incorporate both covalent and noncovalent bonds with the Euclidean distances between the atoms. For this graph neural network, we use two network architectures: 1) a variant of spatial-graph convolutional neural network (SGCNN) used in the potential net (Feinberg et al., 2018) and the fusion methods (Jones et al., 2021), and 2) a type of equivariant graph network architectures, E(n) equivariant graph neural networks (EGNN) (Satorras et al., 2021). Note that although we incorporate SGCNN and EGNN into our MIL framework, any graph neural network architectures can be utilized as a backbone model. We utilize multiple graph neural networks, each of which interprets a single instance containing a docking pose with the protein receptor. We incorporate multiple graph networks into an attention network to capture the docking poses close to the target label.

Given a bag of multiple docking pose instances in a low-dimensional embedding, we propose to use attention-network pooling to predict its single binding affinity. Although docking pose instances are generally ordered by docking scores (with the lowest energy pose listed first), the “best” docking poses (closest to the crystal structures or with the highest correlation to experimental binding affinities) are not necessarily the first ones in the set, according to the pose classification studies (Shim et al., 2022; 2024). To address this and to enable permutation-invariant MIL pooling, we use an attention-based weighted average pooling approach, similar to Luong et al. (2015), Ilse et al. (2018). In this model, the weights for each pose instance are determined by the proposed attention network, allowing for order-independent predictions.

The attention network consists of an attention layer, followed by three fully-connected layers. The attention layer serves as the MIL pooling mechanism: z=k=1Kakhk where ak=σ(tanh(WhkT)). It uses weight parameters (matrix) W, which are learned during training, and applies the hyperbolic tangent function tanh() element-wise. Then it applies the softmax function σ(z)=ezij=1Kezj to ensure that attention scores are interpretable and well-normalized. Note that we also experimented with a multi-head attention-based mechanism (Vaswani et al., 2017). However, no obvious change in prediction accuracy was observed. Figure 1 illustrates our proposed approach.

Figure 1
www.frontiersin.org

Figure 1. Overview of the proposed approach.

3 Experiment setup

In this section, we describe the experimental setup to demonstrate the effectiveness of our proposed approach. We first detail the datasets and the process for generating molecular docking poses. We provide comprehensive details on the training and evaluation procedures, together with the metrics for quantitative assessment.

3.1 Data

The PDBbind database (Su et al., 2019) is the most widely used dataset in structure-based machine learning studies. It combines experimental binding affinity data (logKi or logKd) with curated co-crystal complex structures, originally derived from the Protein Data Bank (PDB) (Burley et al., 2019). Among several versions of the PDBbind database, we use the PDBbind 2020 release, which comprises 19,443 protein-ligand co-crystal structures. The PDBbind database is divided into three subsets: general, refined, and core. The general set is the largest collection of the database with a broad range of complexes. The refined set comprises co-crystal complexes of better quality, curated based on resolution, binding data, and the nature of the complexes. The core set, referred to as CASF-2016, represents a smaller group of high-quality complexes meant to be representative of different protein families and serves as a benchmark for validating docking scoring and machine learning prediction methods. While the general and refined sets are used to train our proposed models, the CASF-2016 (core set) is used to evaluate our methods.

The PDBbind database provides co-crystal complex structures. Thus, we generate molecular docking poses for each of these crystal structures to demonstrate our MIL-based approach. We employed AutoDock VINA (Trott and Olson, 2010; Eberhardt et al., 2021) to generate docking poses by utilizing the ligand and receptor data with the corresponding binding site. Particularly, we used ConveyorLC (Zhang et al., 2014), a VINA docking pipeline for high-performance computing (HPC) clusters. We generated up to 10 poses for each docking ligand. We also computed the root mean square deviation (RMSD) of atomic positions between each docking pose and its corresponding crystal ligand structure using Open Drug Discovery Toolkit (ODDT) (Wójcikowski et al., 2015). Note that AutoDock VINA computes two RMSD values (lower and upper bounds) relative to the first pose, which is the top-ranked pose with the lowest score, rather than using the crystal structure as a reference. Therefore, we used the ODDT software to independently calculate RMSD values based on the crystal structures. The RMSD values serve as metrics for assessing prediction performance across various structural conformations and poses. Figure 2 shows several examples of docking poses and their RMSDs. In docking poses for the PDBbind dataset, there are cases where all poses are close to the crystal structure (poses with RMSDs below 2 Å), as shown in the top example (“1sqa”) of Figure 2. In contrast, sometimes all poses are inaccurate (poses with RMSDs exceeding 4 Å) (“1pxn”). Another case is that the majority of poses are inaccurate whereas a few pose instances are accurate (“3g0w”). Figure 3 shows the overall RMSD errors of the docking poses in the PDBbind dataset.

Figure 2
www.frontiersin.org

Figure 2. Examples of crystal structures and their docking poses of PDBbind CASF-2016 (core set). Autodock VINA (Trott and Olson, 2010; Eberhardt et al., 2021) with ConveyorLC (Zhang et al., 2014) was used for molecular docking. () indicates the RMSD error between the crystal structure and the docking pose.

Figure 3
www.frontiersin.org

Figure 3. Statistics of the docking poses for training and evaluation. The plots represent the RMSD histograms between the docking poses and original crystal ligands in PDBbind 2020 general (left), refined (middle), and CASF-2016 (right) sets. Each histogram bin represents the number of poses within the specific RMSD range.

To evaluate our method on cases where co-crystal structures were not available, we applied our method to an additional compound dataset with molecular docking poses. We selected compound collections targeting the SARS-CoV-2 main protease receptor (“Mpro”) curated by Shim et al. (2024), aimed at identifying potential antiviral agents against the virus. The ligand SMILES data with experimental binding affinities were originally sourced from POSTERA (Morris et al., 2021) and GOSTAR (GoStar, 2023). The receptor structure was obtained from the protein data bank (pdbid: 6LU7) (Burley et al., 2019). This dataset does not provide complex crystal structures. Instead, molecular docking poses were generated using ConveyorLC (Zhang et al., 2014) for AutoDock VINA (Trott and Olson, 2010; Eberhardt et al., 2021). For each compound, we generated up to 10 docking poses. For more detail about the docking process of the Mpro dataset, refer to the original paper (Shim et al., 2024).

We used two different splitting methods to divide the entire Mpro compound set into training and test data: random split and scaffold-based split. The scaffold split ensures that similar compound structures are not present in both the training and test sets by grouping the same Bemis-Murckold scaffolds into either the training or testing set. We utilized the scaffold-based method in the ATOM Modeling PipeLine (AMPL) (Minnich et al., 2020) to generate the training and test split. The number of training and test compounds is 2,135, 1,281, respectively.

3.2 Training and evaluation details

We trained and evaluated our proposed models using PyTorch and PyTorch Geometric. Leveraging their distributed data parallel features, we utilized a cluster of 4 NVIDIA Titan Xp GPUs, each of which has 12 GB of memory. There are two options to train our MIL models: 1) simultaneous training: train both the backbone graph neural network and the MIL attention network together in a unified process, 2) separate training: first, train the backbone models independently and freeze the model parameters to extract features (low-dimensional embedding for 3D complexed pose structures). These features are then used to train the MIL attention network separately. For simultaneous training, we used a mini-batch size of 1 due to GPU memory constraints. While this approach can offer slightly improved predictions for some data, it is computationally expensive because it processes multiple docking pose instances (around 10 poses) with the backbone graph networks simultaneously. In contrast, the separate training option is more computationally efficient. By training the MIL attention network independently, we significantly reduced GPU resource usage and achieved more efficient computation. We used option 2 (separate training) for our experiments.

We used the RMSprop optimizer with the multi-step learning rate scheduler. The total number of epochs is 100-200, depending on the convergence of the loss function. The mini-batch size is 20. Starting with an initial learning rate of 0.001, it decreases by a factor of 0.9 at epoch intervals of 5, 10, 20, 30, 50, 70, and 90. We chose not to utilize checkpoint selection based on the validation split. Instead, we simply opted for the latest checkpoint available for our evaluation.

The hyperparameters for the SGCNN model are as follows: the thresholds for covalent and non-covalent neighbors are set to 1.5 and 4.5 Å, respectively. The size of the gated graph sequence for both covalent and non-covalent bonds is 2. The “gather” widths for covalent and non-covalent bonds are 16 and 12, respectively. We used the following hyperparameters for the EGNN model: the number of edge features is 1, and the model has 6 layers. The distance cut-off is set to 5 Å. Residual connections are used together with attention layers.

4 Results

We performed experiments to demonstrate the effectiveness of the proposed MIL approach using the molecular docking pose structures. Our evaluation report on the performance of our proposed approach with other existing methods used two datasets with their Autodock VINA docking poses. We present the following quantitative metrics: root mean square error (RMSE), mean absolute error (MAE), coefficient of determination (r2), Pearson correlation coefficient (r), and Spearman rank correlation coefficient (ρ).

We benchmarked our results with those from other recent 3D structure-based methods: 1) SGCNN, a variant of Potential Net (Feinberg et al., 2018; Jones et al., 2021), 2) EGNN, equivariant graph neural network, and 3) HACNET (Kyro et al., 2023), which incorporates hybrid attentions through 3DCNN and GCN. We chose SGCNN and EGNN for comparative analysis because our method integrates the architecture into the MIL framework as backbone models. This choice allows us to analyze the behavior of our proposed MIL mechanism. Additionally, we utilized the model checkpoint for HACNET trained on crystal structures in the PDBbind 2016 general and refined sets, due to unavailability of model re-training using the PDBbind 2020 dataset. These comparative models predict binding affinities based on individual docking pose instances. One could report the quantitative metrics using all individual pose results for the comparative models. However, this method does not align with our MIL approach, which generates a single binding affinity from multiple docking poses for each complex entity. Instead, we use the following two methods. First, we report the metrics based on the predicted binding affinities for the top poses, specifically the first pose with the lowest docking scores (“Top”). This approach is commonly used in high-throughput screening with molecular docking tools. Second, we group the results based on compound IDs (e.g., PDB IDs in the PDBbind dataset), and use the average of predicted binding affinities for the same complex entity to ensure consistency with our results (“Avg”).

4.1 PDBbind

Table 1 summarizes the model performance on docking poses with the crystal structures in the PDBbind CASF-2016 (core set). Each complex entity can have up to 11 pose instances, including the crystal ligand instance with up to 10 docking poses. The first group (row 1–6) in Table 1 presents results from models trained exclusively on crystal structures. The SGCNN and EGNN models (row 3–6) were trained using crystal structures from the PDBbind 2020 general and refined sets. In contrast, we utilized the HACNET model (row 1–2) trained on crystal structures from the PDBbind 2016 general and refined sets. The second and third groups (row 7-9 and 10-12, respectively) show results from models trained on both crystal structures and docking poses from the PDBbind 2020 general and refined sets. “Top” presents results based on the top poses, while “Avg” shows results based on non-weighted averages across multiple docking poses within the same complex entities. The SGCNN-MIL and EGNN-MIL results present our proposed MIL method integrating the spatial graph network (SGCNN) and the equivariant graph network (EGNN) as backbone models, respectively.

Table 1
www.frontiersin.org

Table 1. Model Performance of Binding Affinity Prediction on docking and crystal structures of the PDBbind CASF-2016. The models in the first group (row 1–6) were trained solely on the crystal structures of the PDBbind general and refined sets. HACNET was trained using the PDBbind 2016 dataset, while the other models were trained on the PDBbind 2020 dataset. The models in the second and third groups (row 7-9, and row 10–12) were trained on both docking and crystal structures of the PDBbind general and refined sets.

We conducted an additional experiment to show the model performance based training solely on the docking poses without the crystal ligand structures from the PDBbind CASF-2016 (core set). Table 2 shows the model performance among several methods. Similar to the results in Table 1, the first 7 rows present results from HACNET, SGCNN, EGNN, and late-fusion models, respectively, trained exclusively on crystal structures. The HACNET and fusion models (row 1,2 and 7) were trained on crystal structures from the PDBbind 2016 general and refined sets. In contrast, the SGCNN and EGNN models (row 3,4,5 and 6) were trained on crystal structures from the PDBbind 2020 general and refined sets. The second and third groups (row 8-10 and 11-13, respectively) show results from models trained exclusively on docking poses from the PDBbind 2020 general and refined sets. For each of the two backbone models (SGCNN and EGNN), we report the top-pose only (“Top”), averaged (“Avg”) across the poses, and our (“MIL”) results.

Table 2
www.frontiersin.org

Table 2. Model Performance of Binding Affinity Prediction exclusively on docking structures of the PDBbind CASF-2016. The training sets used for the first group follow the same setup as in the previous experiment (Table 1). However, models in the second and third groups were trained solely on docking poses, without crystal structures.

Results in Tables 1, 2 show that our proposed method (“MIL”) outperforms the other methods in both cases. Overall, models trained on both crystal structures and docking poses, or solely on docking poses, outperform those trained exclusively on crystal structures. Among the SGCNN-based models (second group), our MIL method significantly outperforms the other approaches. The results from the third group (EGNN) also demonstrate that our method outperforms the other approaches.

4.2 SARS-CoV-2 Mpro

We conducted experiments using the SARS-CoV-2 (COVID-19) main protease dataset (“Mpro”), sourced from POSTERA (Morris et al., 2021) and GOSTAR (GoStar, 2023), and curated and docking processed by Shim et al. (2024). Our evaluations include both random splits and scaffold splits of the “Mpro” dataset. Tables 3, 4 summarize the model performance in predicting binding affinities on the dataset with random and scaffold splits, respectively.

Table 3
www.frontiersin.org

Table 3. Model Performance of Binding Affinity Prediction on SARS-CoV-2 “Mpro” dataset with a random split.

Table 4
www.frontiersin.org

Table 4. Model Performance of Binding Affinity Prediction on SARS-CoV-2 “Mpro” dataset with a scaffold split.

The evaluation results with the random split, as shown in Table 3, indicate that our proposed method generally outperforms the approaches using only the top docking pose (“Top”) or the averaged pose (“Avg”) results. However, it does not achieve the same level of performance as observed in the PDBbind evaluations. In the scaffold split scenario, where no similar compounds overlap between the training and evaluation sets, our approach clearly outperforms the other methods.

5 Ablation study and discussion

Our experiments with the PDBbind and Mpro datasets demonstrate that our proposed MIL approach consistently outperforms existing methods. We discuss several key observations from these experiments. First, for each backbone architecture, our MIL approach produces more accurate predictions than the non-weighted averages of the predicted binding affinities from multiple docking pose instances and the predictions from the top-ranked poses. The results demonstrate that the attention network module of our approach effectively predicts accurate binding affinities across multiple pose instances within each complex entity. In comparing the EGNN-MIL models trained on both the crystal and docking pose data with those trained on docking pose data alone, there is no significant discrepancy, as illustrated in Tables 1, 2). The model trained on both crystal and docking data exhibits slightly improved RMSE, MAE, and R2 values, while the model trained solely on docking data shows marginal improvements in Pearson and Spearman correlations.

Second, incorporating docking poses generally improve the model prediction over training models solely on crystal structure data. For instance, SGCNN-Top and SGCNN-Avg (rows 7-8 vs. row 3–4) and EGNN-Top and EGNN-Avg (rows 10-11 vs. row 5–6) in Table 2). demonstrate this clearly. It can be argued that incorporating docking data introduces greater diversity into the data distribution, enhancing the model’s predictive performance and generalizability. In addition, adding crystal structures into the training data yields a slight improvement in model accuracy, as shown in a comparison of the results in Table 1 with those in Table 2. However, results may vary based on the quality of the docking poses for each complex structure entity. If all docking poses are significantly inaccurate relative to the crystal structures, incorporating crystal structures could enhance model accuracy.

Furthermore, the choice of backbone graph network architectures also affects the prediction accuracy. The EGNN models demonstrate a clear advantage over the SGCNN models, likely due to its effective feature learning capabilities, which enhance understanding of the input geometry of ligand-protein structures. This results in improved generalizability across different configurations (poses) within the same underlying co-complex entity.

The results evaluating the SARS-CoV-2 Mpro dataset indicate that our MIL approach achieves improved accuracy, compared to the other methods. However, the difference between the correlation coefficients (Pearson and Spearman) becomes marginal in the random split scenario. In contrast, our approach clearly outperforms other methods in the scaffold split scenario. In comparison to the results for the PDBbind dataset, the models trained on the Mpro dataset yield relatively poor performance. It is possibly due to greater inaccuracies in the Mpro docking poses compared to those in the PDBbind dataset.

5.1 Performance with good poses

We evaluate model performance based on the number of good docking poses within each complex entity. We define “good” docking poses as those with an RMSD less than 2.5 Å. We categorize the evaluation data into three groups: structures with no good poses, those with 1–4 good poses, and those with more than 4 good poses. Figure 4 shows model performance based on the number of good poses in the PDBbind CASF-2016 set. Compared to models that rely on top poses (Top) or average predictions across multiple poses (Avg), our approach consistently demonstrates improved and consistent performance. Notably, groups with a sufficient number of good poses tend to outperform those lacking good poses.

Figure 4
www.frontiersin.org

Figure 4. Model performance of binding affinity prediction on the PDBbind CASF-2016 dataset, based on the number of good docking poses within each complex entity. The evaluation data were divided into three groups: structures with no good poses, those with 1–4 good poses, and those with more than 4 good poses. The top and bottom row show results using the SGCNN and EGNN backbones, respectively. The left and right columns show RMSE and R2 results, respectively.

As we discussed earlier, models trained on the Mpro dataset generally underperform, compared to those trained on the PDBbind dataset. One possible explanation is that the quality of the docking poses (the number of “good” poses for each complex entity) may be lower in the Mpro dataset than in the PDBbind dataset. However, analyzing the quality of the docking poses in the Mpro dataset is non-trivial, due to the absence of co-crystal structures. Instead, we measure the RMSD errors between pairs of docking poses within each complex entity. Figure 5 shows the histograms of RMSD errors for pairwise ligand poses in both datasets. While these histograms do not reflect accurate poses based on crystal structures, they indicate the consistency of the docking poses for each complex. The docking poses in the Mpro dataset noticeably show greater discrepancies compared to those in the PDBbind CASF-2016 set. In real compound screening, compounds with highly inconsistent docking poses may be excluded from further ML predictions. Moreover, utilizing more advanced, state-of-the-art molecular docking simulations could improve overall prediction accuracy together with our MIL approach.

Figure 5
www.frontiersin.org

Figure 5. Histograms (probability density) of pair-wise RMSD errors between ligand docking poses in the PDBbind CASF-2016 (left) and SARS-CoV-2 Main Protease (right) datasets.

5.2 Optimal number of docking poses

In multi-instance learning, finding an optimal range of the bag size K (the number of instances within a bag) is crucial. Larger bag sizes generally yield better feature extraction and improved model accuracy by capturing diverse instance characteristics. In contrast, larger bag sizes can increase the number of model parameters and overall complexity, which may lead to overfitting. Having more positive and negative instances can also introduce label ambiguity during model training. In our MIL setup, we generated up to 10 docking poses in most cases. We evaluated our models using varying numbers of docking pose instances for each complex entry, specifically K = 1, 3, 5, 7, and 9 instances. During model evaluation, we randomly selected K instances from 10 poses and averaged the predictions over 10 trials. Figure 6 illustrates the RMSE and R2 values of our MIL models with two different backbone architectures. Overall, utilizing more docking poses tends to increase model accuracy. Specifically, SGCNN-MIL model achieves the best performance with 7 docking poses, while EGNN-MIL model performs optimally with 5 and 9 docking poses. While determining an optimal bag size is challenging due to factors such as compound type, the number of good poses, and model architectures, a range of 5–10 poses per bag seems to be more effective, compared to smaller bag sizes.

Figure 6
www.frontiersin.org

Figure 6. Model performance of binding affinity prediction on the PDBbind CASF-2016 dataset, based on the number of docking pose instances within each complex entity. The evaluation was performed where the number of poses is 1, 3, 5, 7, and 9. The top and bottom row show results using the SGCNN and EGNN backbones, respectively. The left and right columns show RMSE and R2 results, respectively. Error bars indicate minimum and maximum values.

5.3 Attention types

We evaluated two different attention types for multi-instance learning. The first method involves global attention for weighted average pooling, as detailed in Section 2. The second method utilizes a multi-head attention mechanism (Vaswani et al., 2017). We observed no significant change or improvement using the multi-head attention network. While the multi-head attention mechanism offers slight improvements in predictions for SGCNN-MIL models (RMSE, MAE and R2), it yields poorer results for EGNN-MIL models, as illustrated in Table 5.

Table 5
www.frontiersin.org

Table 5. Model Performance in Binding Affinity Prediction on PDBbind CASF-2016 using two types of attention mechanisms.

5.4 Computational Costs

The attention network in our MIL framework is lightweight, comprising approximately 68,000 parameters, which is under 1 MB in size. The training and evaluation of our models are fast. Using a single NVIDIA Titan Xp GPU, the training time for one epoch with a batch size of 20, across 18,539 complex entities with up to 10 docking poses, is approximately 36.5 s (1.97 milliseconds per complex entity).

6 Conclusion

We presented a new structure-based multi-instance learning approach utilizing molecular docking poses as a bag of complex entity. Our method employs an attention network together with graph neural networks to enable permutation-invariant weighted average pooling between docking poses. We demonstrated our method using the PDBbind and the Mpro datasets. Our method offers binding affinity prediction without requiring co-crystal structures, which increases its applicability for various targets of interest.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

HK: Conceptualization, Data curation, Formal Analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing–original draft, Writing–review and editing. HS: Data curation, Investigation, Writing–review and editing, Software. AR: Software, Writing–review and editing. SH: Resources, Software, Writing–review and editing. GS: Resources, Writing–review and editing. JA: Funding acquisition, Project administration, Supervision, Writing–review and editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was funded by the Defense Threat Reduction Agency (DTRA), HDTRA1242044 (HK, SH, GS) and HDTRA1036045 (HS, AR, JA).

Acknowledgments

This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under contract DE-AC52-07NA27344. Lawrence Livermore National Security, LLC. LLNL-JRNL-2000673.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Backenköhler, M., Groß, J., Wolf, V., and Volkamer, A. (2024). Guided docking as a data generation approach facilitates structure-based machine learning on kinases. J. Chem. Inf. Model. 64, 4009–4020. doi:10.1021/acs.jcim.4c00055

PubMed Abstract | CrossRef Full Text | Google Scholar

Burley, S. K., Berman, H. M., Bhikadiya, C., Bi, C., Chen, L., Costanzo, L. D., et al. (2019). Protein data bank: the single global archive for 3d macromolecular structure data. Nucleic Acids Res. 47, D520–D528. doi:10.1093/nar/gky949

PubMed Abstract | CrossRef Full Text | Google Scholar

Carbonneau, M.-A., Cheplygina, V., Granger, E., and Gagnon, G. (2018). Multiple instance learning: a survey of problem characteristics and applications. Pattern Recognit. 77, 329–353. doi:10.1016/j.patcog.2017.10.009

CrossRef Full Text | Google Scholar

Chithrananda, S., Grand, G., and Ramsundar, B. (2020). Chemberta: large-scale self-supervised pretraining for molecular property prediction. Corr. abs/2010, 09885. doi:10.48550/arXiv.2010.09885

CrossRef Full Text | Google Scholar

Corso, G., Stärk, H., Jing, B., Barzilay, R., and Jaakkola, T. (2023). “Diffdock: diffusion steps, twists, and turns for molecular docking,” in International conference on learning representations (ICLR).

Google Scholar

Dooly, D. R., Zhang, Q., Goldman, S. A., and Amar, R. A. (2002). Multiple instance learning of real valued data. J. Mach. Learn. Res. 3, 651–678.

Google Scholar

Duvenaud, D., Maclaurin, D., Aguilera-Iparraguirre, J., Gómez-Bombarelli, R., Hirzel, T., Aspuru-Guzik, A., et al. (2015). “Convolutional networks on graphs for learning molecular fingerprints,” in Proceedings of the 28th international conference on neural information processing systems - volume 2 (Cambridge, MA, USA: MIT Press), NIPS’15), 2224–2232.

Google Scholar

Eberhardt, J., Santos-Martins, D., Tillack, A. F., and Forli, S. (2021). Autodock vina 1.2.0: new docking methods, expanded force field, and python bindings. J. Chem. Inf. Model. 61, 3891–3898. doi:10.1021/acs.jcim.1c00203

PubMed Abstract | CrossRef Full Text | Google Scholar

Fatima, S., Ali, S., and Kim, H.-C. (2023). A comprehensive review on multiple instance learning. Electronics 12, 4323. doi:10.3390/electronics12204323

CrossRef Full Text | Google Scholar

Feinberg, E. N., Sheridan, R., Joshi, E., Pande, V. S., and Cheng, A. C. (2019). Step change improvement in admet prediction with potentialnet deep featurization. arXiv:1903. doi:10.48550/arXiv.1903.11789

CrossRef Full Text | Google Scholar

Feinberg, E. N., Sur, D., Wu, Z., Husic, B. E., Mai, H., Li, Y., et al. (2018). Potentialnet for molecular property prediction. ACS Cent. Sci. 4, 1520–1530. doi:10.1021/acscentsci.8b00507

PubMed Abstract | CrossRef Full Text | Google Scholar

Friesner, R. A., Banks, J. L., Murphy, R. B., Halgren, T. A., Klicic, J. J., Mainz, D. T., et al. (2004). Glide: a new approach for rapid, accurate docking and scoring. 1. method and assessment of docking accuracy. J. Med. Chem. 47, 1739–1749. doi:10.1021/jm0306430

PubMed Abstract | CrossRef Full Text | Google Scholar

Fuchs, F. B., Worrall, D. E., Fischer, V., and Welling, M. (2020). “Se(3)-transformers: 3d roto-translation equivariant attention networks,” in Proceedings of the 34th international conference on neural information processing systems (NIPS’20: Curran Associates Inc).

Google Scholar

Gomes, J., Ramsundar, B., Feinberg, E., and Pande, V. (2017). Atomic convolutional networks for predicting protein-ligand binding affinity.

Google Scholar

GoStar (2023). Global online structure activity relationship database (gostar).

Google Scholar

Halgren, T. A., Murphy, R. B., Friesner, R. A., Beard, H. S., Frye, L. L., Pollard, W. T., et al. (2004). Glide: a new approach for rapid, accurate docking and scoring. 2. enrichment factors in database screening. J. Med. Chem. 47, 1750–1759. doi:10.1021/jm030644s

PubMed Abstract | CrossRef Full Text | Google Scholar

Ilse, M., Tomczak, J. M., and Welling, M. (2018). Attention-based deep multiple instance learning. arXiv Prepr. arXiv:1802.04712. doi:10.48550/arXiv.1802.04712

CrossRef Full Text | Google Scholar

Jiang, D., Hsieh, C.-Y., Wu, Z., Kang, Y., Wang, J., Wang, E., et al. (2021). Interactiongraphnet: a novel and efficient deep graph representation learning framework for accurate protein–ligand interaction predictions. J. Med. Chem. 64, 18209–18232. doi:10.1021/acs.jmedchem.1c01830

PubMed Abstract | CrossRef Full Text | Google Scholar

Jiménez, J., Doerr, S., Martínez-Rosell, G., Rose, A. S., and De Fabritiis, G. (2017). Deepsite: protein-binding site predictor using 3d-convolutional neural networks. Bioinformatics 33, 3036–3042. doi:10.1093/bioinformatics/btx350

PubMed Abstract | CrossRef Full Text | Google Scholar

Jiménez, J., Škalič, M., Martínez-Rosell, G., and De Fabritiis, G. (2018). Kdeep: protein-ligand absolute binding affinity prediction via 3d-convolutional neural networks. J. Chem. Inf. Model. 58, 287–296. doi:10.1021/acs.jcim.7b00650

PubMed Abstract | CrossRef Full Text | Google Scholar

Jones, D., Kim, H., Zhang, X., Zemla, A., Stevenson, G., Bennett, W. F. D., et al. (2021). Improved protein–ligand binding affinity prediction with structure-based deep fusion inference. J. Chem. Inf. Model. 61, 1583–1592. doi:10.1021/acs.jcim.0c01306

PubMed Abstract | CrossRef Full Text | Google Scholar

Kyro, G. W., Brent, R. I., and Batista, V. S. (2023). Hac-net: a hybrid attention-based convolutional neural network for highly accurate protein-ligand binding affinity prediction. J. Chem. Inf. Model. 63, 1947–1960. doi:10.1021/acs.jcim.3c00251

PubMed Abstract | CrossRef Full Text | Google Scholar

Lu, W., Wu, Q., Zhang, J., Rao, J., Li, C., and Zheng, S. (2022). “Tankbind: trigonometry-aware neural networks for drug-protein binding structure prediction,” in Proceedings of the 36th international conference on neural information processing systems.

Google Scholar

Luong, T., Pham, H., and Manning, C. D. (2015). “Effective approaches to attention-based neural machine translation,” in Proceedings of the 2015 conference on empirical methods in natural language processing, 1412–1421. doi:10.18653/v1/D15-1166

CrossRef Full Text | Google Scholar

Mazuz, E., Shtar, G., Kutsky, N., Rokach, L., and Shapira, B. (2023). Pretrained transformer models for predicting the withdrawal of drugs from the market. Bioinformatics 39, btad519. doi:10.1093/bioinformatics/btad519

PubMed Abstract | CrossRef Full Text | Google Scholar

McNutt, A., Francoeur, P., Aggarwal, R., Masuda, T., Meli, R., Ragoza, M., et al. (2021). Gnina 1.0: molecular docking with deep learning. J. Cheminformatics 13, 43. doi:10.1186/s13321-021-00522-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Méndez-Lucio, O., Ahmad, M., Rio-Chanona, E. A. d., and Wegner, J. K. (2021). A geometric deep learning approach to predict binding conformations of bioactive molecules. Nat. Mach. Intell. 3, 1033—–1039. doi:10.1038/s42256-021-00409-9

CrossRef Full Text | Google Scholar

Minnich, A. J., McLoughlin, K., Tse, M., Deng, J., Weber, A., Murad, N., et al. (2020). Ampl: a data-driven modeling pipeline for drug discovery. J. Chem. Inf. Model. 60, 1955–1968. doi:10.1021/acs.jcim.9b01053

PubMed Abstract | CrossRef Full Text | Google Scholar

Moriwaki, H., Tian, Y.-S., Kawashita, N., and Takagi, T. (2018). Mordred: a molecular descriptor calculator. J. Cheminf 10, 4. doi:10.1186/s13321-018-0258-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Morris, A., McCorkindale, W., Consortium, T., Drayman, N., Chodera, J., Tay, S., et al. (2021). Discovery of sars-cov-2 main protease inhibitors using a synthesis-directed de novo design model. Chem. Commun. 57, 5909–5912. doi:10.1039/d1cc00050k

PubMed Abstract | CrossRef Full Text | Google Scholar

Mqawass, G., and Popov, P. (2024). graphlambda: fusion graph neural networks for binding affinity prediction. J. Chem. Inf. Model. 64, 2323–2330. doi:10.1021/acs.jcim.3c00771

PubMed Abstract | CrossRef Full Text | Google Scholar

O’Boyle, N. M., Banck, M., James, C. A., Morley, C., Vandermeersch, T., and Hutchison, G. R. (2011). Open babel: an open chemical toolbox. J. Cheminf. 3, 33. doi:10.1186/1758-2946-3-33

PubMed Abstract | CrossRef Full Text | Google Scholar

Powers, A. S., Yu, H. H., Suriana, P., and Dror, R. O. (2022). Fragment-based ligand generation guided by geometric deep learning on protein-ligand structure. bioRxiv. doi:10.1101/2022.03.17.484653

CrossRef Full Text | Google Scholar

Ragoza, M., Hochuli, J., Idrobo, E., Sunseri, J., and Koes, D. R. (2017). Protein–ligand scoring with convolutional neural networks. J. Chem. Inf. Model. 57, 942–957. doi:10.1021/acs.jcim.6b00740

PubMed Abstract | CrossRef Full Text | Google Scholar

Ramsundar, B., Eastman, P., Walters, P., Pande, V., Leswing, K., and Wu, Z. (2019). Deep Learning for the life Sciences (O’reilly media). Available at: https://www.amazon.com/Deep-Learning-Life-Sciences-Microscopy/dp/1492039837.

Google Scholar

RDKit (2024). Rdkit: open-source cheminformatics.

Google Scholar

Rogers, D., and Hahn, M. (2010). Extended-connectivity fingerprints. J. Chem. Inf. Model. 50, 742–754. doi:10.1021/ci100050t

PubMed Abstract | CrossRef Full Text | Google Scholar

Rong, Y., Bian, Y., Xu, T., Xie, W., Wei, Y., Huang, W., et al. (2020). Self-supervised graph transformer on large-scale molecular data. Adv. Neural Inf. Process. Syst. 33. doi:10.48550/arXiv.2007.02835

CrossRef Full Text | Google Scholar

Satorras, V. G., Hoogeboom, E., and Welling, M. (2021). “E(n) equivariant graph neural networks,” in Proceedings of the 38th international conference on machine learning. (PMLR), vol. 139 of Proceedings of machine learning research. Editors M. Meila, and T. Zhang 9323–9332.

Google Scholar

Scantlebury, J., Vost, L., Carbery, A., Hadfield, T. E., Turnbull, O. M., Brown, N., et al. (2023). A small step toward generalizability: training a machine learning scoring function for structure-based virtual screening. J. Chem. Inf. Model. 63, 2960–2974. doi:10.1021/acs.jcim.3c00322

PubMed Abstract | CrossRef Full Text | Google Scholar

Shim, H., Allen, J. E., and Bennett, W. F. D. (2024). Enhancing docking accuracy with pecan2, a 3d atomic neural network trained without co-complex crystal structures. Mach. Learn. Knowl. Extr. 6, 642–657. doi:10.3390/make6010030

CrossRef Full Text | Google Scholar

Shim, H., Kim, H., Allen, J. E., and Wulff, H. (2022). Pose classification using three-dimensional atomic structure-based neural networks applied to ion channel–ligand docking. J. Chem. Inf. Model. 62, 2301–2315. doi:10.1021/acs.jcim.1c01510

PubMed Abstract | CrossRef Full Text | Google Scholar

Stärk, H., Beaini, D., Corso, G., Tossou, P., Dallago, C., Günnemann, S., et al. (2022a). “3d infomax improves gnns for molecular property prediction,” in Proceedings of the 39th international Conference on machine learning. Vol. 162 of Proceedings of machine learning research, 20479–20502.

Google Scholar

Stärk, H., Ganea, O., Pattanaik, L., Barzilay, R., and Jaakkola, T. (2022b). “Equibind: geometric deep learning for drug binding structure prediction,” in International Conference on machine learning (PMLR), 20503–20521.

Google Scholar

Stepniewska-Dziubinska, M. M., Zielenkiewicz, P., and Siedlecki, P. (2018). Development and evaluation of a deep learning model for protein–ligand binding affinity prediction. Bioinformatics 34, 3666–3674. doi:10.1093/bioinformatics/bty374

PubMed Abstract | CrossRef Full Text | Google Scholar

Su, M., Yang, Q., Du, Y., Feng, G., Liu, Z., Li, Y., et al. (2019). Comparative assessment of scoring functions: the casf-2016 update. J. Chem. Inf. Model. 59, 895–913. doi:10.1021/acs.jcim.8b00545

PubMed Abstract | CrossRef Full Text | Google Scholar

tfbio (2018). tfbio python package

Google Scholar

Trott, O., and Olson, A. J. (2010). Autodock vina: improving the speed and accuracy of docking with a new scoring function, efficient optimization and multithreading. J. Comput. Chem. 31, 455–461. doi:10.1002/jcc.21334

PubMed Abstract | CrossRef Full Text | Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2017). “Attention is all you need,”. Advances in neural information processing systems. Editors I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathanet al. (NY, United States: Curran Associates, Inc.), 30.

Google Scholar

Wójcikowski, M., Zielenkiewicz, P., and Siedlecki, P. (2015). Open drug discovery toolkit (oddt): a new open-source player in the drug discovery field. J. Cheminformatics 7. 26. doi:10.1186/s13321-015-0078-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, H., Liao, L., Saravanan, K. M., Yin, P., and Wei, Y. (2019). Deepbindrg: a deep learning based method for estimating effective protein-ligand affinity. PeerJ 7, e7362. doi:10.7717/peerj.7362

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, S., Jin, Y., Liu, T., Wang, Q., Zhang, Z., Zhao, S., et al. (2023). Ss-gnn: a simple-structured graph neural network for affinity prediction. ACS Omega 8, 22496–22507. doi:10.1021/acsomega.3c00085

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, X., Wong, S. E., and Lightstone, F. C. (2014). Toward fully automated high performance computing drug discovery: a massively parallel virtual screening pipeline for docking and molecular mechanics/generalized born surface area rescoring to improve enrichment. J. Chem. Inf. Model. 54, 324–337. doi:10.1021/ci4005145

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhou, G., Gao, Z., Ding, Q., Zheng, H., Xu, H., Wei, Z., et al. (2023). “Uni-mol: a universal 3d molecular representation learning framework,” in International conference on learning representations. doi:10.26434/chemrxiv-2022-jjm0j-v3

CrossRef Full Text | Google Scholar

Keywords: AI-driven drug development, virtual high-throughput screening, protein-ligand interaction, molecular docking, 3D atomic graph representation, structure-based machine learning, multi-instance learning, attention mechanism

Citation: Kim H, Shim H, Ranganath A, He S, Stevenson G and Allen JE (2025) Protein-ligand binding affinity prediction using multi-instance learning with docking structures. Front. Pharmacol. 15:1518875. doi: 10.3389/fphar.2024.1518875

Received: 29 October 2024; Accepted: 26 November 2024;
Published: 03 January 2025.

Edited by:

Moom Rahman Roosan, Chapman University, United States

Reviewed by:

Larisa Ivanova, University of Tartu, Estonia
Mayasah Al-Nema, UCSI University, Malaysia

Copyright © 2025 Kim, Shim, Ranganath, He, Stevenson and Allen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Hyojin Kim, aGtpbUBsbG5sLmdvdg==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.