Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Edge-enhanced interaction graph network for protein-ligand binding affinity prediction

Abstract

Protein-ligand interactions are crucial in drug discovery. Accurately predicting protein-ligand binding affinity is essential for screening potential drugs. Graph neural networks have proven highly effective in modeling spatial relationships and three-dimensional structures within intermolecular. In this paper, we introduce a graph neural network-based model named EIGN to predict protein-ligand binding affinity. The model consists of three main components: the normalized adaptive encoder, the molecular information propagation module, and the output module. Experimental results indicate that EIGN achieves root mean squared error of 1.126 and Pearson correlation coefficient of 0.861 on CASF-2016. Additionally, our model outperforms state-of-the-art methods on CASF-2013, CASF-2016, and the CSAR-NRC set, showing exceptional accuracy and robust generalization ability. To further validate the effectiveness of EIGN, we conducted several experiments, including ablation studies, feature importance analysis, data similarity analysis, and others, to evaluate its performance and applicability.

Introduction

Protein-ligand interactions form the foundation of numerous physiological processes within organisms, playing a crucial role in understanding drug mechanisms, discovering novel drugs, and designing targeted therapeutic strategies [13]. Accurately predicting Protein-Ligand binding Affinity (PLA) is an essential step in identifying potential drug candidates during the drug development process [46]. However, protein-ligand interactions are highly complex and influenced by multiple factors such as molecular structure, interaction sites, and molecular dynamics, which makes affinity prediction more challenging [7,8].

Early methods for affinity prediction primarily relied on traditional experimental techniques, including X-ray crystallography [9] and nuclear magnetic resonance [10]. While these methods deliver fairly accurate affinity data, their high time and cost requirements make them impractical for efficiently analyzing large compound libraries in the early stages of drug screening [1113]. Consequently, computational methods have increasingly become a research hotspot for affinity prediction [14]. Computer-aided drug design methods, such as molecular docking [15], molecular dynamics simulations [16], and deep learning models [17], provide new insights and tools for rapidly predicting PLA [18,19]. Among these, deep learning models stand out for their remarkable potential, as evidenced by their success in natural language generation [20], image generation [21], and speech detection [22].

In the past, some deep learning methods for affinity prediction have utilized the sequence information of proteins and ligands. These methods feed the sequence data into deep learning models, using sequence-based architectures such as Long Short-Term Memory (LSTM) networks, Convolutional Neural Networks (CNNs), or attention-based Transformer models to predict affinity [23,24]. For example, DeepDTAF [25] employs dilated convolutions to capture long-range interactions and integrates both local features of the protein binding pocket and global features of the entire protein to predict binding affinity. CAPLA [26] leverages cross-attention mechanisms to capture interaction features between protein binding pockets and ligands. By analyzing amino acid sequences or molecular formulas, these models effectively identify specific patterns and local structural features at the sequence level, thereby enabling them to achieve promising performance in predicting binding affinity. However, sequence-based models often struggle with capturing spatial relationships and complex three-dimensional structures. This limitation hampers their ability to capture subtle structural variations and detailed molecular interactions in realistic environments [27].

Graph Neural Networks (GNNs) have emerged in recent years as a powerful tool for modeling the intricate relationships in molecular structures, gaining widespread use in biomolecular studies [28,29]. By representing proteins and ligands as molecular graphs with nodes and edges, GNNs can capture inter-molecular interaction information through graph convolution operations, thereby enabling more precise representations for affinity prediction [3032]. SIGN [33] introduced a polarity-guided graph attention layer that uses the distance and angle information between atoms. FAST [34] integrates local and global contextual features to predict PLA. IGN [35] represents protein-ligand complexes as ligand graphs, protein graphs, and bimolecular protein-ligand graphs. GIGN [36] applies a heterogeneous interaction layer to unify covalent and non-covalent interactions within the message-passing stage, enabling it to learn node representation more effectively. CurvAGN [37] introduced curvature blocks and adaptively guided neural blocks to encode edge attributes in biomolecular graphs. LGN [38] enhances the capture of local and global features in protein-ligand complexes by incorporating extra ligand feature extraction. PIGNet [39] employs a physics-informed GNN, which performs excellently in scoring and screening tasks. PIGNet2 [40] further enhances model generalizability by expanding training data with a broader range of binding conformations and ligand types. These GNN-based methods perform well in affinity prediction. However, most existing models typically combine inter- and intra-molecular interactions, which may limit their ability to capture local structural details. Although some studies have attempted to handle these two types of interactions separately, the methods remain relatively simplistic, making it challenging for models to extract rich local information and accurately capture complex molecular interactions.

In this paper, we propose a GNN-based method for predicting PLA. The method focuses on refining the modeling of interactions within protein-ligand complexes by utilizing the inter- and intra-molecular message-passing modules. To enhance the representational power of edge features for capturing interaction information between nodes better, we employ an edge update mechanism that integrates node feature information into edge features. This design helps the model leverage edge information to update node features effectively during message passing. We evaluate the performance of the model on the test set and compare it with state-of-the-art methods. The results demonstrate the superior performance of EIGN. Additionally, several experiments were conducted to evaluate the model in a comprehensive manner, such as ablation studies, feature importance analysis, data similarity analysis, and others.

Materials and methods

Database

In this study, we utilized the PDBbind v2020 database as the primary data source. PDBbind [41] is a database of high-quality protein-ligand complexes, designed for structural biology and computational chemistry. It is extensively utilized in drug discovery and molecular modeling research [42]. Compiled from the Protein Data Bank, this database provides a comprehensive collection of protein-ligand complexes, offering extensive data support for studying protein-ligand interactions.

The PDBbind database provides structural information of complexes along with their experimentally measured binding affinities (e.g., dissociation constant Kd, inhibition constant Ki), which form a reliable foundation for building and validating PLA prediction models. PDBbind v2020 includes a greater number of updated complex data compared to earlier versions, significantly enhancing the comprehensiveness and accuracy of the database. Specifically, PDBbind v2020 (N = 19,443) was randomly divided into two subsets: one subset as the model training set (N = 16,954) and the other subset as the validation set (N = 2,000). During the dataset partitioning, we excluded samples that overlapped with the test set as well as samples that could not be processed by RDKit. The test set consisted of CASF-2013 (N = 195) and CASF-2016 (N = 285) from the PDBbind database [43,44].

In addition to the PDBbind database, we also used a portion of the CSAR-NRC set (N = 85) as an additional test set to further evaluate the generalization capability and predictive performance of the model. The CSAR-NRC set is one of the datasets released by the Community Structure-Activity Resource (CSAR) to provide high-quality PLA data for validating and comparing various virtual screening and molecular docking models [45]. This dataset includes a rigorously filtered set of protein-ligand complexes with precisely measured binding affinities, making it one of the standard datasets for assessing affinity prediction models.

Protein–ligand complex representation

Graph structures are particularly well-suited for representing protein-ligand complexes, as they effectively capture atomic interactions and the topological information within the complex [46]. Interactions between proteins and ligands (such as drugs) occur at the active sites of proteins, which are some depressions and cavities, usually known as binding pockets [47]. In this study, we defined the protein-ligand interaction region using a distance threshold of 5.0 Å, considering only residues within this range around the ligand. This threshold was selected based on previous studies [48], balancing prediction accuracy and computational cost.

To represent these interactions, we constructed two types of graphs: one for inter-molecular interactions and another for intra-molecular interactions. In these graph structures, each node represents an atom, and we used six types of information to construct the features of these nodes. These features were processed through one-hot encoding and transformed into vector representations. The edges between nodes were represented using either Euclidean distance or node degree.

A notable aspect of our approach is the unified representation method for both protein and ligand atoms. We applied the same representation method without additional feature distinctions, ensuring generality and scalability. The specific details of node features are summarized in Table 1.

In addition to the mentioned feature processing, we introduced an edge augmentation strategy to better handle intermolecular interactions. Specifically, during graph construction, we randomly delete certain edges to simulate structural noise that may result from docking errors or data quality issues. For instance, we removed edges between nodes at greater distances (e.g., edges exceeding 4 Å). Meanwhile, the edge enhancement strategy randomly adds new edges to enrich the diversity of the graph, allowing the model to contact different graph structures during the learning process. The three types of graphs used in EIGN are illustrated in Fig 1.

Model architecture

The framework of the EIGN model is shown in Fig 2A. The input of the model is the protein-ligand complex. We divide the overall model into three modules: Normalized Adaptive Encoder (NAE), the molecular message propagation module, and the output module. The first two modules primarily use GNN to extract and update the feature representations of the nodes. The node features updated by the first two modules are passed to the output Module to generate the final prediction.

thumbnail
Fig 2. Model architecture.

(A) Overall framework of EIGN. (B) Inter-molecular message passing structure.

https://doi.org/10.1371/journal.pone.0320465.g002

NAE.

NAE is an encoder module that integrates linear transformations of graph features, a graph propagation mechanism, and global pooling operations. To capture higher-order relationships between nodes, we introduced an enhanced graph propagation mechanism—Approximate Personalized Propagation of Neural Predictions (APPNP) [49]. APPNP employs a “skip connection” gating mechanism that uses the hyperparameter α to regulate the weighted combination of initial and propagated features. This mechanism, compared to traditional GNNs, ensures that the influence of the initial features is retained at every layer, thereby effectively mitigating the over-smoothing problem. Consequently, APPNP enables NAE to capture long-range dependencies more effectively and enhances its performance on molecular graph tasks. To ensure numerical stability during the information propagation process, we apply L2 normalization to the node features before propagation. The normalized features are then amplified by a scaling factor, improving their ability to convey information during propagation. These normalized features are passed into the propagation module for further integration of the structural information of the graph. Unlike traditional GNNs, our final step incorporates the node features X after propagation and the graph-level embedding obtained through sum pooling, as the result of NAE. The following formula illustrates the execution flow of NAE:

(1)(2)(3)

where H represents the output after feature transformation. W is the weight matrix for the linear transformation. X denotes the propagated node features. s is the scaling factor, set to 1.8 in this study. α represents the propagation rate, set to 0.1. P is the normalized adjacency matrix. is the graph-level embedding. represents the set of all nodes in graph i.

Molecular message propagation module.

The molecular message propagation module is divided into three parts: inter-molecular message passing, intra-molecular message passing, and the edge perturbation layer. The first two components handle inter-molecular and intra-molecular interactions, while the edge perturbation layer processes the inter-molecular interactions after edge perturbation.

In the inter-molecular message passing module, we designed a mechanism that combines different types of graph convolutional layers to capture the complex structural features within molecular graphs. Fig 2B illustrates the detailed structure of this module. This module updates the node features using both GINE and DGCN, and the final output is obtained by concatenating the node features updated by each method. The GINE layer is an enhanced variant of the Graph Isomorphism Network (GIN) [50]. It allows edge features to be used as input and dynamically integrates them during message passing to improve node feature updates. In this study, edge features processed by the edge update mechanism are fed into the GINE layer to further boost the expressive power of the node features. DGCN module enriches node feature representations with a multi-stage message passing strategy. It first applies a Graph Convolutional Network (GCN) to process the input node features and extract local structural information. Next, the feature update process introduced an edge weighting mechanism based on node degree, which propagates and adjusts node features globally through a custom message passing layer. The resulting features from both parts are concatenated as the output of the DGCN module, hence effectively combining local and global features. The node feature update process in the DGCN module can be described as follows:

(4)(5)(6)

where represents the set of neighboring nodes of node i. is the number of neighbors of node i. denotes the node features of neighboring node j.

The primary function edge update mechanism is to dynamically update the edge features in the graph by combining the feature information of two nodes, enhancing the model’s ability to represent node relationships. By integrating node features and processing them through a Multi-Layer Perceptron (MLP), the mechanism refines the feature representation of each edge, thereby enhancing the ability of the graph structure to capture intricate relationships. This mechanism captures detailed interaction patterns between nodes, offering richer edge information to the GNN, which ultimately improves the effectiveness of graph data modeling. The workflow of the edge update mechanism is illustrated in Fig 2B.

For intra-molecular message passing, the module structure is similar to inter-molecular message passing, but it focuses on the node and edge features within the molecule. Since the internal structure of the molecule contains unique local structural features, intra-molecular message passing helps capture micro-interactions within the molecule by focusing on the local flow of information within the molecular nodes.

The edge perturbation layer addresses inter-molecular interactions influenced by edge perturbations. It aims to enhance the diversity and robustness of the graph structure, enabling the capture of rich inter-node relationships, especially those that may emerge as potential structural information after edge addition or removal. Unlike the previous two sub-modules, this module uses only GIN for graph processing and does not utilize edge feature inputs.

Finally, in the molecular message propagation module, the node features updated through the three sub-modules are combined, and the combined features are used as the final output of the module.

Output module.

As shown in Fig 2A, the output Module mainly consists of an MLP and a pooling layer. Its role is to convert node-level features into a global graph-level representation through fully connected layers, pooling, and fusion operations, which are then used for predicting PLA. What makes this module unique is its approach of separately integrating the graph-level embeddings from both the NAE and molecular message propagation module. This enables the model to capture the interactions between local and global features, ultimately improving its ability to model complex structural relationships.

Implementation details

The model training was conducted on a single GPU, with a batch size of 128 and a dropout rate of 0.1. For optimization, the Adam optimizer was employed, utilizing a learning rate of 5e-4 and a weight decay of 1e-6. The training was conducted for 800 epochs with early stopping implemented, terminating the training process if there was no improvement in performance on the validation set for 100 consecutive epochs. The model used the mean squared error loss function as the objective to evaluate the regression performance.

Results and discussion

This section comprehensively evaluates the performance of the EIGN model in PLA prediction using two metrics: Root Mean Squared Error (RMSE) and Pearson correlation coefficient (Rp). The experiments show that EIGN outperforms state-of-the-art methods on the PDBbind core set and CSAR-HIQ-set, demonstrating strong generalization ability and prediction accuracy. Ablation studies confirm the effectiveness of key modules within the model, while analyses of feature contribution and similarity-based sample ratios further identify the crucial factors influencing model performance. In addition, we evaluated the performance of the model in virtual screening tasks and its capability to handle unknown target and ligand structures.

Experiments on the PDBbind core set

We used three datasets to train the model: the refined set of PDBbind v2020 (N = 4473), the general set with refined samples removed (N = 12585), and the complete general set (N = 16954). Each training set was repeated three times, each with a different random seed. The mean and standard deviation of the three model predictions were used as performance indicators. The experimental results are shown in Table 2.

As shown in Table 2, the model trained on the refined set and the model trained on the general set with the refined samples removed demonstrate similar performance on both test sets. However, the general set contains more data noise, which leads to more fluctuation in the prediction results, as evidenced by the larger standard deviations of RMSE and Rp shown in Table 2. The complete general set strikes a balance in terms of data scale and diversity, leading to the best generalization ability for the trained model.

Fig 3 presents the scatter plots of affinity prediction and actual affinity values for EIGN trained on the PDBbind v2020 general set, evaluated on the validation set, CASF-2013, and CASF-2016. From the scatter plot of the validation set, the model predictions generally align with the actual values. However, in the extreme value regions, the model performs poorly. This issue may arise from the scarcity of extreme samples and the imbalance in the data, which prevents the model from accurately predicting the affinity values for these cases. The predicted and actual values for the two test sets also show a clear linear relationship.

thumbnail
Fig 3. Scatter plots of predicted values (y-axis) versus actual values (x-axis) for EIGN on the validation set, CASF-2013, and CASF-2016.

https://doi.org/10.1371/journal.pone.0320465.g003

Comparison with state-of-the-art methods

To verify the superiority of EIGN, we compared its performance against several state-of-the-art methods on CASF-2013 and CASF-2016. To ensure a fair comparison, we limited the evaluation to methods that utilized either the PDBbind v2016 or PDBbind v2020 datasets for training. A total of seven methods were selected for this comparison: OnionNet [51], SIGN [33], FAST [34], IGN [35], GIGN [36], CurvAGN [37], and LGN [38]. Specifically, we reproduced IGN and GIGN and trained them on the same datasets used for EIGN.

Table 3 presents the results of this comparison. The results highlight that our model consistently outperforms the other methods. Regardless of whether it was trained on the PDBbind v2016 or PDBbind v2020 datasets, EIGN demonstrates superior performance and surpasses all other methods in the comparison. It is important to note that although some methods use the same version of the dataset (e.g., PDBbind v2016 or PDBbind v2020), the number of training samples varies. This discrepancy arises from differences in preprocessing and filtering criteria employed in different studies. For instance, the GIGN method excludes samples that cannot be processed using RDKit. In contrast, CurvAGN does not filter any samples. Additionally, the significant difference in the number of samples in CASF-2013 is due to some methods only using the subset of samples that overlap between CASF-2013 and CASF-2016, rather than the complete CASF-2013 sample set.

thumbnail
Table 3. Comparison of predicted results between EIGN and state-of-the-art methods on CASF-2013 and CASF-2016.

https://doi.org/10.1371/journal.pone.0320465.t003

For the model trained on PDBbind v2016, EIGN outperforms LGN on CASF-2016 with a 4% decrease in RMSE and a 2% increase in Rp. When trained on PDBbind v2020, EIGN achieves a 12% reduction in RMSE and a 6% improvement in Rp compared to GIGN on CASF-2016.

Ablation study

To evaluate the effectiveness of each component in our model, we created five variants of EIGN:

  1. A. Replacing NAE with MLP.
  2. B. Removing edge update mechanism.
  3. C. Removing edge perturbation layer.
  4. D. Using only GINE to handle intra- and inter-molecular interactions.
  5. E. Applying the gating mechanism from NAE to subsequent modules.

We trained these variant models and assessed their predictive performance on CASF-2013 and CASF-2016, enabling us to analyze the contribution of each component to the overall model performance. A comparison of the prediction results for the different variants on two datasets is shown in Fig 4.

thumbnail
Fig 4. Prediction performance of EIGN and four variants on CASF-2013 and CASF-2016.

https://doi.org/10.1371/journal.pone.0320465.g004

From Fig 4, it can be observed that replacing NAE with MLP has the greatest impact on the performance of the model. This modification significantly increased the RMSE and decreased Rp, indicating that NAE plays a crucial role in capturing node features and the global information of the graph. The other modules contribute supplementarily to edge feature updates, structural noise handling, and local information representation. Maintaining these modules helps to improve the generalization ability and stability of the model in chemical structure data.

Performance evaluation on the non-PDBbind external test Set

To validate the generalization capability of EIGN and its predictive performance on non-PDBbind datasets, we used the CSAR-NRC dataset as an external test set for performance evaluation. Using the same training set and strategy, we trained two interaction GNNs, GIGN and IGN, and compared their prediction performance on the CSAR-NRC dataset with that of our model. As shown in Fig 5, EIGN outperforms GIGN and IGN on the CSAR-HIQ-set, demonstrating higher prediction accuracy and good generalization ability.

thumbnail
Fig 5. Comparison of prediction performance of EIGN, GIGN, and IGN on the CSAR-HIQ-set.

https://doi.org/10.1371/journal.pone.0320465.g005

Analysis of node feature contribution

To further analyze the contribution of different node features to the performance of EIGN, we conducted ablation studies and feature importance analysis.

In the ablation study, we sequentially removed different node features, retrained the model, and evaluated its performance on the test set to observe the impact of each feature. The experimental results are shown in Fig 6A. The removal of atomic symbols and implicit valence features has a significant impact on model performance, indicating that these features play a crucial role in the prediction task.

thumbnail
Fig 6. Contribution analysis of the node feature components in the model.

(A) Ablation study. (B) GNNExplainer.

https://doi.org/10.1371/journal.pone.0320465.g006

Additionally, we used GNNExplainer [52] to analyze the importance of node features. The procedure involved loading the trained best model, interpreting it on specific samples, and generating a feature importance map. This map illustrates the contribution of each feature in the model. The explanation results are shown in Fig 6B. Except for aromatic features, there was no significant difference in the importance of various feature groups, suggesting that aromatic features have a minor influence on model performance.

From Fig 6, it can be observed that aromatic features consistently have the smallest impact on model performance, which indicates that aromatic information contributes little to the EIGN. However, the impact of removing atomic degree features in the ablation study was minor, which differs from the results in GNNExplainer. This discrepancy may be due to our model already adopting degree-related edge features, hence the removal of degree features did not result in a significant decline in model performance.

Impact of dataset similarity on prediction results

We analyzed the impact of dataset similarity on prediction results. Specifically, we evaluated the structural similarity between proteins in the PDBbind v2020 refined set and those in CASF2016. Similarity was quantified using the TM-score [53], a widely used metric for structural alignment. Samples with a TM-score greater than 0.5 were considered similar.

Based on this criterion, after removing samples that overlap with CASF2016, we obtained 1,734 similar samples and 3,236 remaining samples from the PDBbind v2020 refined set. Using these samples, we constructed six training sets, each containing 3,000 samples, with the proportions of similar samples set at 0%, 10%, 20%, 30%, 40%, and 50%. We trained EIGN using these training sets and evaluated its prediction performance on CASF2016. As shown in Fig 7, the model achieves better performance on the test set when the training set contains a higher proportion of similar samples. For example, when the proportion of similar samples increases from 0% to 50%, the prediction accuracy of the model rises substantially. This observation suggests that incorporating a more diverse range of protein structures in the training data is crucial for achieving more accurate predictions. Therefore, future data collection efforts should prioritize covering a variety of protein structures, especially those currently underrepresented or missing in the training set. This approach will enhance the generalization performance and predictive capability of the model.

thumbnail
Fig 7. The impact of the proportion of similar samples in the training set on model prediction performance.

https://doi.org/10.1371/journal.pone.0320465.g007

Performance of the model in virtual screening tasks

After training the model on the PDBbind dataset, we further evaluated its virtual screening capabilities. Specifically, we focused on the PYGM target from the DUD-E dataset and conducted ligand docking analyses using both experimental structures and those predicted by AlphaFold3 [54]. To assess the reliability of the predicted structure, we performed a global structural alignment with TM-align [55]. The alignment produced a TM-score of 0.711—well above the 0.5 threshold—confirming that both structures share the same fold. We then concentrated on the active pocket region, defined as the 108 residues located within 5Å of the ligand in the experimental structure, and observed a local Root Mean Square Deviation (RMSD) of 0.47Å. Based on these validation results, we carried out restricted molecular docking with AutoDock Vina, setting the search space as a 20Å cube centered on the active pocket.

To evaluate the prediction results, we conducted a visual analysis of the outcomes for both experimental setups. Density plots and ROC curves were generated, as shown in Fig 8.

thumbnail
Fig 8. Visualization of the performance of EIGN, GIGN, and IGN on the PYGM target.

(A) Experimentally resolved protein structure. (B) Alphafold3-predicted protein structure.

https://doi.org/10.1371/journal.pone.0320465.g008

Analysis of the density plots for the experimentally resolved protein structure indicates that the EIGN model excels in distinguishing active compounds from inactive ones. The ROC curve analysis further supports this, with the EIGN model achieving an AUC value of 0.85, outperforming the other two models.

For the AlphaFold3-predicted protein target, although the EIGN model still surpassed the other models, there was a noticeable decline in its overall performance. This may be attributed to the precision limitations of the AlphaFold3 predicted structures and deviations in the active site region.

Prediction ability of the model for ligands with undetermined structures

To evaluate the prediction ability of the model for ligands with undetermined structures, we conducted experimental analysis on the Cyclin-Dependent Kinase 2 (CDK2) target from the ChEMBL database. The structure of each ligand was generated using the RDKit tool based on the Simplified Molecular Input Line Entry System (SMILES) representation of the ligands.

As shown in Fig 9, the prediction results of EIGN, GIGN, and IGN for this target were visualized in scatter plots. The results indicate that for target CDK2, the predictions from EIGN, GIGN, and IGN were suboptimal, as the predicted affinity values showed minimal variation across multiple ligands for the same target.

thumbnail
Fig 9. Scatter plots of the prediction results of EIGN, GIGN, and IGN on the CDK2 target.

https://doi.org/10.1371/journal.pone.0320465.g009

The poor performance of the three models on the CDK2 target may be closely related to the characteristics of the PDBbind dataset. The PDBbind dataset consists of protein-ligand complexes, where each protein is typically associated with a single high-resolution ligand. This structure trains models to recognize the specific features of a protein interacting with a single high-affinity ligand, rather than handling scenarios where a protein is associated with multiple ligands. In contrast, the ChEMBL dataset is more diverse; a single target is often associated with numerous ligands, exhibiting a broader range of binding affinities. Additionally, the ligands for the CDK2 target display a high degree of similarity, which might make it challenging for the model to differentiate subtle variations among these ligands.

Although the dataset characteristics largely determine the performance of the model, the model itself may also have limitations. EIGN primarily extracts graph structural features through the GNN layers when processing protein-ligand complexes. In cases where multiple potential ligands exist for the same target, the model struggles to distinguish subtle structural differences or variations in affinity between ligands. For example, the message passing and feature aggregation in GNN layers tend to smooth node features, reducing sensitivity to local structural differences in ligands.

Future work could address these limitations by incorporating multi-source training, such as integrating data from ChEMBL or other multi-ligand target datasets during model training. Additionally, the ability of the model to capture ligand differences could be enhanced by introducing additional constraints or attention mechanisms to better focus on ligand similarities, such as molecular fingerprints or topological features.

Conclusion

In this study, we proposed an innovative GNN-based model for PLA prediction, and it is named EIGN. This model captures inter- and intra-molecular interactions through a multi-level message-passing mechanism within GCN. Our model integrates inter- and intra-molecular message-passing strategies to better capture diverse interaction patterns. Additionally, we incorporate an edge update mechanism to enhance the representational capacity of edge features, enabling a more comprehensive modeling of relationships within the graph.

To verify the performance of the model, we performed evaluations on high-quality datasets such as CASF-2013, CASF-2016, and CSAR-NRC set. Based on the evaluation indicators RMSE and Rp, the experimental results show that EIGN exhibits significant performance advantages in PLA prediction.

Despite the significant performance improvements, the model still has potential for further enhancement, particularly in screening capabilities. Future research should focus on optimizing this aspect to augment its practical applicability.

Acknowledgments

We are grateful for technical support from the High Performance Computing Platform of Xiangtan University.

References

  1. 1. Metz J, Hajduk P. Rational approaches to targeted polypharmacology: Creating and navigating protein–ligand interaction networks. Curr Opin Chem Biol. 2010;14(4):498–504.
  2. 2. Vangone A, Bonvin AMJJ. Prodigy: A contact-based predictor of binding affinity in protein-protein complexes. Bio Protoc. 2017;7(3):e2124. pmid:34458447
  3. 3. Cheng T, Li Q, Zhou Z, Wang Y, Bryant SH. Structure-based virtual screening for drug discovery: A problem-centric review. AAPS J. 2012;14(1):133–41. pmid:22281989
  4. 4. Wang Y, Guo Y, Kuang Q, Pu X, Ji Y, Zhang Z, et al. A comparative study of family-specific protein-ligand complex affinity prediction based on random forest approach. J Comput Aided Mol Des. 2015;29(4):349–60. pmid:25527073
  5. 5. Thafar M, Raies A, Albaradei S, Essack M, Bajic V. Comparison study of computational prediction tools for drug-target binding affinities. Fic. 2019;7:782.
  6. 6. Chen X, Yan CC, Zhang X, Zhang X, Dai F, Yin J, et al. Drug-target interaction prediction: Databases, web servers and computational models. Brief Bioinform. 2016;17(4):696–712. pmid:26283676
  7. 7. Kitchen D, Decornez H, Furr J, Bajorath J. Docking and scoring in virtual screening for drug discovery: Methods and applications. Nat Rev Drug Disco. 2004;3(11):935–49.
  8. 8. De Ruyck J, Brysbaert G, Blossey R, Lensink M. Molecular docking as a popular tool in drug design, an in silico travel. Bioinformatics Ai, Chem. 2016;1–11.
  9. 9. McPherson A, Gavira J. Introduction to protein crystallization. J Appl Crystallogr. 2014;70(1):2–20.
  10. 10. Miyanishi K, Sugiki T, Matsui T, Ozawa R, Hatanaka Y, Enozawa H, et al. Protein-ligand interaction analyses with nuclear magnetic resonance spectroscopy enhanced by dissolution triplet dynamic nuclear polarization. J Phys Chem Lett. 2023;14(27):6241–7. pmid:37401781
  11. 11. Dhakal A, McKay C, Tanner JJ, Cheng J. Artificial intelligence in the prediction of protein-ligand interactions: Recent advances and future directions. Brief Bioinform. 2022;23(1):bbab476. pmid:34849575
  12. 12. Korlepara D, Vasavi C, Jeurkar S, Pal P, Roy S, Mehta S, et al. Plas-5k: Dataset of protein-ligand affinities from molecular dynamics for machine learning applications. 2022;9(1):548.
  13. 13. N Politis S, Colombo P, Colombo G, M Rekkas D. Design of experiments (DoE) in pharmaceutical development. Drug Dev Ind Pharm. 2017;43(6):889–901. pmid:28166428
  14. 14. Kyro GW, Brent RI, Batista VS. HAC-Net: a hybrid attention-based convolutional neural network for highly accurate protein-ligand binding affinity prediction. J Chem Inf Model. 2023;63(7):1947–60. pmid:36988912
  15. 15. Wu Q, Peng Z, Zhang Y, Yang J. COACH-D: improved protein-ligand binding sites prediction with refined ligand-binding poses through molecular docking. Nucleic Acids Res. 2018;46(W1):W438–42. pmid:29846643
  16. 16. Wang DD, Ou-Yang L, Xie H, Zhu M, Yan H. Predicting the impacts of mutations on protein-ligand binding affinity based on molecular dynamics simulations and machine learning methods. Comput Struct Biotechnol J. 2020;18:439–54. pmid:32153730
  17. 17. Karimi M, Wu D, Wang Z, Shen Y. DeepAffinity: interpretable deep learning of compound-protein affinity through unified recurrent and convolutional neural networks. Bioinformatics. 2019;35(18):3329–38. pmid:30768156
  18. 18. Meli R, Morris GM, Biggin PC. Scoring functions for protein-ligand binding affinity prediction using structure-based deep learning: A review. Front Bioinform. 2022;2:885983. pmid:36187180
  19. 19. Wang H. Prediction of protein-ligand binding affinity via deep learning models. Brief Bioinform. 2024;25(2):bbae081. pmid:38446737
  20. 20. Achiam J, Adler S, Agarwal S, Ahmad L, Akkaya I, Aleman F, et al. Gpt-4 technical report. 2023.
  21. 21. Borji AJapa. Generated faces in the wild: Quantitative comparison of stable diffusion, midjourney and dall-e 2. 2022.
  22. 22. Harar P, Alonso-Hernandez J, Mekyska J, Galaz Z, Burget R, Smekal Z, et al. Voice pathology detection using deep learning: a preliminary study. International Conference and Workshop on Bioinspired Intelligence (IWOBI). IEEE; 2017.
  23. 23. Abbasi K, Razzaghi P, Poso A, Amanlou M, Ghasemi JB, Masoudi-Nejad A. DeepCDA: deep cross-domain compound-protein affinity prediction through LSTM and convolutional neural networks. Bioinformatics. 2020;36(17):4633–42. pmid:32462178
  24. 24. Monteiro NRC, Oliveira JL, Arrais JP. DTITR: End-to-end drug-target binding affinity prediction with transformers. Comput Biol Med. 2022;147:105772. pmid:35777085
  25. 25. Wang K, Zhou R, Li Y, Li M. DeepDTAF: a deep learning method to predict protein-ligand binding affinity. Brief Bioinform. 2021;22(5):bbab072. pmid:33834190
  26. 26. Jin Z, Wu T, Chen T, Pan D, Wang X, Xie J, et al. CAPLA: improved prediction of protein-ligand binding affinity by a deep learning approach based on a cross-attention mechanism. Bioinformatics. 2023;39(2):btad049. pmid:36688724
  27. 27. Jumper J, Evans R, Pritzel A, Green T, Figurnov M, Ronneberger O, et al. Highly accurate protein structure prediction with AlphaFold. Nature. 2021;596(7873):583–9. pmid:34265844
  28. 28. Harren T, Gutermuth T, Grebner C, Hessler G, Rarey M. Modern machine‐learning for binding affinity estimation of protein–ligand complexes: Progress, opportunities, and challenges. WIRCMS. 2024;14(3):e1716.
  29. 29. Gainza P, Sverrisson F, Monti F, Rodolà E, Boscaini D, Bronstein MM, et al. Deciphering interaction fingerprints from protein molecular surfaces using geometric deep learning. Nat Methods. 2020;17(2):184–92. pmid:31819266
  30. 30. Gilmer J, Schoenholz SS, Riley PF, Vinyals O, Dahl GE, editors. Neural message passing for quantum chemistry. International conference on machine learning; 2017.
  31. 31. Torng W, Altman RB. Graph Convolutional Neural Networks for Predicting Drug-Target Interactions. J Chem Inf Model. 2019;59(10):4131–49. pmid:31580672
  32. 32. Fout A, Byrd J, Shariat B, Ben-Hur AJAinips. Protein interface prediction using graph convolutional networks. 2017;30.
  33. 33. Li S, Zhou J, Xu T, Huang L, Wang F, Xiong H, editors. Structure-aware interactive graph neural networks for the prediction of protein-ligand binding affinity. Proceedings of the 27th ACM SIGKDD conference on knowledge discovery & data mining. 2021.
  34. 34. Jones D, Kim H, Zhang X, Zemla A, Stevenson G, Bennett WFD, et al. Improved Protein-Ligand Binding Affinity Prediction with Structure-Based Deep Fusion Inference. J Chem Inf Model. 2021;61(4):1583–92. pmid:33754707
  35. 35. Jiang D, Hsieh C-Y, Wu Z, Kang Y, Wang J, Wang E, et al. InteractionGraphNet: A Novel and Efficient Deep Graph Representation Learning Framework for Accurate Protein-Ligand Interaction Predictions. J Med Chem. 2021;64(24):18209–32. pmid:34878785
  36. 36. Yang Z, Zhong W, Lv Q, Dong T, Yu-Chian Chen C. Geometric Interaction Graph Neural Network for Predicting Protein-Ligand Binding Affinities from 3D Structures (GIGN). J Phys Chem Lett. 2023;14(8):2020–33. pmid:36794930
  37. 37. Wu J, Chen H, Cheng M, Xiong H. CurvAGN: Curvature-based Adaptive Graph Neural Networks for Predicting Protein-Ligand Binding Affinity. J Bioinform Comput Biol. 2023;24(1):378.
  38. 38. Guo J. Improving structure-based protein-ligand affinity prediction by graph representation learning and ensemble learning. PLoS One. 2024;19(1):e0296676. pmid:38232063
  39. 39. Moon S, Zhung W, Yang S, Lim J, Kim WY. PIGNet: a physics-informed deep learning model toward generalized drug-target interaction predictions. Chem Sci. 2022;13(13):3661–73. pmid:35432900
  40. 40. Moon S, Hwang S-Y, Lim J, Kim W. PIGNet2: a versatile deep learning-based protein–ligand interaction prediction model for binding affinity scoring and virtual screening. Journal of Drug Design and Discovery. 2024;3(2):287–99.
  41. 41. Wang R, Fang X, Lu Y, Wang S. The PDBbind database: collection of binding affinities for protein-ligand complexes with known three-dimensional structures. J Med Chem. 2004;47(12):2977–80. pmid:15163179
  42. 42. Liu Z, Li Y, Han L, Li J, Liu J, Zhao Z, et al. PDB-wide collection of binding data: current status of the PDBbind database. Bioinformatics. 2015;31(3):405–12. pmid:25301850
  43. 43. Su M, Yang Q, Du Y, Feng G, Liu Z, Li Y, et al. Comparative assessment of scoring functions: The casf-2016 update. J Chem Inf Model. 2019;59(2):895–913. pmid:30481020
  44. 44. Li Y, Su M, Liu Z, Li J, Liu J, Han L, et al. Assessing protein-ligand interaction scoring functions with the CASF-2013 benchmark. Nat Protoc. 2018;13(4):666–80. pmid:29517771
  45. 45. Dunbar JB Jr, Smith RD, Yang C-Y, Ung PM-U, Lexa KW, Khazanov NA, et al. CSAR benchmark exercise of 2010: Selection of the protein-ligand complexes. J Chem Inf Model. 2011;51(9):2036–46. pmid:21728306
  46. 46. Zitnik M, Leskovec J. Predicting multicellular function through multi-layer tissue networks. Bioinformatics. 2017;33(14):i190–8. pmid:28881986
  47. 47. Zhao J, Cao Y, Zhang L. Exploring the computational methods for protein-ligand binding site prediction. J Comput Biol. 2020;18417–26.
  48. 48. Zhang S, Jin Y, Liu T, Wang Q, Zhang Z, Zhao S, et al. SS-GNN: A Simple-structured graph neural network for affinity prediction. ACS Omega. 2023;8(25):22496–507. pmid:37396234
  49. 49. Gasteiger J, Bojchevski A, Günnemann S. Predict then propagate: Graph neural networks meet personalized pagerank. 2018.
  50. 50. Xu K, Hu W, Leskovec J, Jegelka S. How powerful are graph neural networks? 2018.
  51. 51. Zheng L, Fan J, Mu Y. OnionNet: a Multiple-Layer Intermolecular-Contact-Based Convolutional Neural Network for Protein-Ligand Binding Affinity Prediction. ACS Omega. 2019;4(14):15956–65. pmid:31592466
  52. 52. Ying Z, Bourgeois D, You J, Zitnik M, Leskovec J. Gnnexplainer: Generating explanations for graph neural networks. Adv Neural Inf Process Syst. 2019;32.
  53. 53. Zhang Y, Skolnick J. Scoring function for automated assessment of protein structure template quality. Proteins. 2004;57(4):702–10. pmid:15476259
  54. 54. Abramson J, Adler J, Dunger J, Evans R, Green T, Pritzel A, et al. Accurate structure prediction of biomolecular interactions with AlphaFold 3. 1–3.
  55. 55. Zhang Y, Skolnick J. TM-align: a protein structure alignment algorithm based on the TM-score. Nucleic Acids Res. 2005;33(7):2302–9. pmid:15849316