Figures
Abstract
Biomedical texts provide important data for investigating drug-drug interactions (DDIs) in the field of pharmacovigilance. Although researchers have attempted to investigate DDIs from biomedical texts and predict unknown DDIs, the lack of accurate manual annotations significantly hinders the performance of machine learning algorithms. In this study, a new DDI prediction framework, Subgraph Enhance model, was developed for DDI (SubGE-DDI) to improve the performance of machine learning algorithms. This model uses drug pairs knowledge subgraph information to achieve large-scale plain text prediction without many annotations. This model treats DDI prediction as a multi-class classification problem and predicts the specific DDI type for each drug pair (e.g. Mechanism, Effect, Advise, Interact and Negative). The drug pairs knowledge subgraph was derived from a huge drug knowledge graph containing various public datasets, such as DrugBank, TwoSIDES, OffSIDES, DrugCentral, EntrezeGene, SMPDB (The Small Molecule Pathway Database), CTD (The Comparative Toxicogenomics Database) and SIDER. The SubGE-DDI was evaluated from the public dataset (SemEval-2013 Task 9 dataset) and then compared with other state-of-the-art baselines. SubGE-DDI achieves 83.91% micro F1 score and 84.75% macro F1 score in the test dataset, outperforming the other state-of-the-art baselines. These findings show that the proposed drug pairs knowledge subgraph-assisted model can effectively improve the prediction performance of DDIs from biomedical texts.
Author summary
Drug-drug interactions occur when two or more drugs react with each other, potentially leading to adverse drug reactions that can manifest as toxicity, reduced treatment efficacy, and, in extreme cases, patient mortality. With the growing number of approved drugs, detecting negative drug-drug interactions has become a crucial concern in pharmacovigilance. However, manually mining these interactions from the vast and rapidly expanding pool of published biomedical texts, such as clinical trial reports and literature, is arduous and challenging. To address this issue, various machine learning methods have been developed and applied to text mining for both known and unknown drug-drug interactions. Most of these methods primarily predict interactions by considering only drug molecular properties and text characters, neglecting valuable information about interactions between different entities, such as drugs, pathways, and side effects. In response, our approach seeks to enhance the performance of drug-drug interaction prediction by leveraging knowledge graphs constructed from a variety of biomedical entities. Our experimental results validate this approach, demonstrating that incorporating interactive information from knowledge graphs leads to the development of a more efficient model for predicting drug-drug interactions.
Citation: Shi Y, He M, Chen J, Han F, Cai Y (2024) SubGE-DDI: A new prediction model for drug-drug interaction established through biomedical texts and drug-pairs knowledge subgraph enhancement. PLoS Comput Biol 20(4): e1011989. https://doi.org/10.1371/journal.pcbi.1011989
Editor: James Gallo, University at Buffalo - The State University of New York, UNITED STATES
Received: September 26, 2023; Accepted: March 11, 2024; Published: April 16, 2024
Copyright: © 2024 Shi et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The source code and data are available at https://github.com/syy528000/SubGE-DDI.git.
Funding: This work was supported by the Department of Education of Guangdong Province under Grants No.2020KZDZX1126 (to YC) and the Guangdong Medical Products Administration under Grants No.2022ZDZ06 (to YC and FH). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Drug-drug interaction (DDI) refers to a change in drug’s effects due to the presence of another drug [1]. This commonly occurs in cases of polypharmacy, when the effects of one drug alter the effects of other drugs in a combination regimen. DDI may enhance or weaken the efficacy of the drug, causing adverse drug reactions (ADRs), which can even be life-threatening in severe cases [2]. Some databases, such as DrugBank [3], TWOSIDES [4], DDInter [5], KEGG [6], BIOSNAP [7] and MecDDI [8] have been established to provide information related to interactions between drugs and promote the development of new drugs while avoiding ADRs. Currently, the rapid growth in the number of biomedical publications makes it increasingly difficult to manually extract valuable DDI information from articles, despite its critical importance [9]. This necessitates the development of automated DDI extraction methods.
Task 9 of the 2013 International Symposium on Semantic Evaluation (SemEval-2013), "Extracting drug-drug interactions from biomedical texts" (DDIExtraction 2013), is the benchmark corpus of this domain [10]. In the field of natural language processing (NLP), researchers have proposed several methods for predicting DDIs with biomedical text information [11–13].
Despite the extensive use of traditional neural networks for DDI extraction in biomedical texts before 2018, there is a need for more effective methods as the available models are largely dependent on manual labeling and unsatisfactory prediction accuracy. As a result, BERT model (proposed by Jacob Devlin in 2018) has been widely applied in recent years [14]. Moreover, BERT and its derivative models are employed in DDIs extraction with good results. Chen et al. [15] proposed a novel method based on BioBERT [16] for extracting DDIs without adding any external drug information, where the drug name is not converted into standard tokens. Molina et al. [17] proposed a framework that leverages Gaussian noise injection to enhance the performance of DDI prediction. Besides, some researchers have shown that external drug-related information can further improve the effectiveness of the model [18–19]. For instance, Asada et al. (2021) proposed a new method that can extract DDIs by combining information from external drug databases and large-scale plain text [20,21].
In addition, studies in the field of Knowledge Graph (KG) have shown that KG can effectively integrate multiple entity type and complex relationships between biological entities. These methods improve the extraction of informative high-order semantic features, which enhances the DDI prediction accuracy [22]. Deep learning has significantly enhanced DDI prediction, giving rise to a variety of frameworks that capitalize on different information sources. One approach optimizes the Biomedical Knowledge Graph (BKG) by integrating local and global information, as demonstrated by Ren et al. [23]. Another, proposed by Su et al. [24], leverages an attention-based KG representation learning framework. Meanwhile, Gu et al. [25] employed supervised contrastive learning with DDI data as negative samples to transform drug embedding vectors and predict interactions. Su et al. [26] adopted KG2ECapsule, which utilizes a capsule graph neural network to generate high-quality negative samples for DDI prediction. Tang et al. (2023) [27] propose a novel approach called DSIL-DDI that derives domain-agnostic representations of DDIs from a source domain, enhancing model generalizability and interpretability.
Although previous studies primarily performed DDI extraction from biomedical texts, relying solely on text information and drug molecular features, Duan et al. [28] introduced an improved approach in which molecular structure information is integrated with text, emphasizing functional group structure to enhance the extraction process. Further, He et al. [29] optimized the model’s performance by integrating 3D molecular graph structure and position information. These methods improved the model performance of DDIs extraction using drug molecular structures and achieved good results. However, drug polymorphisms limit DDIs extraction. To address the limitations of current methods used to illustrate chemical structures, we encompass more biomedical entities in DDIs extraction, such as diseases and pathways, providing a better understanding of drug interactions.
In this study, a novel DDIs extraction framework involving external BKG, SubGE-DDI, was proposed to find another strategy to prevent these limitations. This model combines text features and drug pairs-related KG information for DDIs extraction. First, a BKG was built from several public biomedical databases, including DrugBank, TWOSIDES, OFFSIDES [30], SIDER [31], SMPDB [32], DrugCentral [33], Entrez Gene [34], and CTD [35] to generate the KG features of the drug pairs. The message of targeted drug pairs with position embedding was enhanced using CNN. Moreover, three pre-trained models are compared and the most suitable for DDIs extraction is determined. Finally, the pre-trained PubMedBERT-based [36] text feature extraction model was adopted because it can achieve the highest F1 score on both micro-averaged and macro-averaged metric. Moreover, a Subgraph -Attentional Graph Convolutional Network (SubAGCN) method was used to develop KG. SubAGCN can effectively anchor the relevant subgraph of drug pairs in KG and generate inference paths in the subgraph through a novel attention module. Three hidden layer fusion methods were proposed to improve the combination of subgraph features generated by SubAGCN and the text features generated by PubMedBERT. The resultant mixed vectors, including drug pairs KG interactive messages, and text features, are employed to conduct the final classification task. The results show that SubGE-DDI can accurately extract the DDI relationships and enhance macro-averaged and micro-averaged metrics, obtaining F1macro and F1micro of 84.75% and 83.91%, respectively.
Materials and methods
Datasets
Training data were obtained from SemEval-2013 Task 9 dataset [10]. This dataset was utilized to investigate automatic drug recognition and DDIs extraction algorithms based on biomedical texts. Data are obtained from two public datasets, including DrugBank and MedLine. A total of 730 and 175 articles about DDIs were obtained from the DrugBank and MedLine, respectively. The target drugs and interaction types in the aforementioned articles were annotated by experts. Subsequently, the dataset was divided into training and testing sets. Published models for extracting DDIs from biomedical texts are compared in this dataset [15,20,28].
The dataset defines the following four interaction labels and a negative label.
- Mechanism (Mec.): Used to annotate DDIs described by their Pharmacokinetics (PK) mechanism (e.g. Grepafloxacin may inhibit the metabolism of theobromine).
- Effect (Eff.): Used to annotate DDIs describing an effect (e.g. About 46% of uninfected volunteers develop rash while receiving SUSTIVA and clarithromycin) or a Pharmacodynamics (PD) mechanism (e.g. Chlorthalidone may potentiate the action of other antihypertensive drugs).
- Advise (Adv.): Used when giving a recommendation or advice regarding a drug interaction (e.g. UROXATRAL should not be used in combination with other alpha-blockers).
- Int.: Used when a DDI appears in the text without providing any additional information (e.g. The interaction of omeprazole and ketoconazole has been established).
- Negative (Neg.): used when there is no interaction between two drugs.
In summary, DDIs extraction is a multiclass classification task for classifying target entities in input sentences. entity pairs will be generated if there are n entities in an input sentence. Herein, the drug pair was labeled as “Drug1” and “Drug2” to ensure the generalization of features. An example of preprocessing is shown in Table 1.
Most samples in Table 2 are negative pairs, indicating no interaction between the drugs in the sentence. This phenomenon often leads to data imbalance, which affects the performance of DDI extraction models based on machine learning. In this study, the negative instances were filtered as much as possible according to a negative instance filtering strategy. The filtering strategy from previous studies [37–39] is listed below (Table 3).
Rule 1.–Drug pairs that refer to the same name or drugs that are abbreviations of another drug should be screened.
Rule 2.–Drug pairs that are in coordinate structure should be screened.
Rule 3.–A drug that is a special case of the other one should be filtered.
To ensure accurate mapping between drug pairs in the sentences and entities in the knowledge graph, drug names were mapped to their corresponding DrugBank IDs. Any pair without a DrugBank ID was excluded from the analysis. There were many negative pairs in the training and testing sets after filtering (Table 4). Therefore, a multi-focal loss was used to avoid the sample imbalance. Multi-focal loss is a loss function [40] that treats different DDI types as different weights.
Multi-focal loss was determined as follows:
(1)
m, αi, and pt represent the number of DDI types, weight of each DDI type as defined in Eq (2), and prediction of the DDI prediction model, respectively (γ = 2).
(2)
Counti and m represents the number of the i−th type and the number of different types, respectively.
The overview of SubGE-DDI
This study introduces SubGE-DDI, a novel model for extracting DDIs. SubGE-DDI leverages SubAGCN to extract biomedical subgraph knowledge for the corresponding drug pair entity. Text information of DDIs was obtained using the pre-trained PubMedBERT-based text feature extraction model. The message of targeted drug pairs was enhanced by position embedding via CNN. Finally, the three hidden layer fusion methods were compared to determine the best approach for fusing the subgraph features and the text features.
An illustration of the proposed SubGE-DDI model is presented in Fig 1. For example, the sentence “Administration of valproic acid decreases oral clearance of temozolomide by about 5%.” was input where “valproic acid” and “temozolomide” are drug pair. The subgraph information for the drug pair was then obtained through SubAGCN. The positional information of drug pairs, text information, and subgraph information were combined as a fusion feature. Finally, the fusion feature was used as the input for DDI predictions.
The SubGE-DDI framework consists of three key parts–subgraph information, text features and fusion part. The subgraph information section is shown in the figure and the text features section and fusion part is illustrted in (A).
Drug pairs features in knowledge subgraph
The biomedical KG is a comprehensive mechanism that illustrates human biology. For example, other nodes (e.g. pathway, protein, etc.) may change when a drug-drug paired node in the KG changes, resulting in a series of reactions and producing various physiological outcomes. By learning the structural features of the biomedical knowledge graph surrounding targeted drug pairs, we provide important data that can be leveraged to reveal the biological mechanisms underlying their effects on the body. Specifically, graph structure features are located inside the node and edge structures of the network and can be calculated directly from the graph. However, it may take longer and consume a massive amount of RAM if the entire KG is learnt in a computer. Zhang et al. [41] (2018) demonstrated that the local enclosing subgraphs contain enough information to effectively learn the structure features of the entire graph. In this study, we focused on local subgraphs around the drug pairs in the KG and the location information of the nodes within subgraphs when collecting information about drug interactions.
Specifically, the k−hop neighboring nodes were extracted for the target drug pairs u and v, and
. d(∙,∙) represents the distance between two nodes on
. The enclosing subgraph was obtained based on the intersections of these nodes,
. Meanwhile, the node i representation for each node i in the subgraph
, was updated as
(
), where ⊕ is concatenate).
A new GCN-based model, named SubAGCN, was designed to fully learn the structure features of subgraphs, and summarize the subgraph information into a graph-based pathway for potential drug interactions. Notably, the pathway is a sparse subgraph since drug interactions often contain varying amounts of complex interplays among many types of biomedical entities. Therefore, a layer-independent, relationship-aware module with self-attention was designed. The weight for each edge was configured in .
Specifically, the self-attention module consists of two parts: First, the weights should be assigned to each edge in the subgraph since the importance of the relationship between each entity in the subgraph will be different, even if the edge type is the same. Therefore, γ and β (Brockschmidt [42]) were determined to ensure that the model can dynamically weight features based on information at the edge target nodes, as follows:
(3)
where g and
represent a single linear layer and the targeted node representation of the edge in t layer, respectively.
The unimportant edges were removed from the subgraph to keep only important edges. Second, the neighboring nodes in the subgraph were taken into consideration to generate the attention weight and calculate αi,j, which is the signal strength score of the edge between nodes i and j [43,44]. αi,j was calculated as follows:
(4)
(5)
where WI and WJ represent the weights of the individual linear layers corresponding to the source and target nodes, respectively; ri,j represents the relation between node i and node j;
represents the size of the node input vector h(0); Tanh(∙) represents the tanh function for non-linear transformation, which makes αi,j range from -1 to 1. A threshold hyperparameter ζ was designed to screen out unimportant edges by setting their weights τi,j to 0, when the αi,j is lower than ζ.
A key subgraph important for targeted drug pairs was obtained using the self-attention module. The information of subgraph was then integrated using the following message-passing scheme. For each node v, the neighbor node message was determined as follows:
(6)
where
and
represent the set containing all the neighbors of node v in the subgraph and the weight matrix of the relationship r between nodes u, v at the layer t, respectively. Basis factorization was used to decompose
into linear combinations of a small number of basis matrices {Vb}b∈B and avoid overfitting (Schlichtkrull et al. [45]) as follows:
(7)
The biomedical entities associative information of node v was then updated by the neighbor node message
as follows:
(8)
where Wself represents the weight matrix to transform the node embedding itself.
Otherwise, the mean of all node embeddings in subgraph at layer t was used to represent the information of subgraph as follows:
(9)
where WSub and
represent the weight matrix for nodes transformation in subgraph and embedding of the node i at layer t in subgraph, respectively.
Finally, the layer-aggregation mechanism was employed to integrate the various representations generated by each layer [46]. The node/subgraph embeddings in each layer, such as hv and , were concatenated. Finally, the node embeddings of the target drug pairs and the subgraph embeddings were concatenated to obtain the drug pairs representation hdp.
Drug pairs features in text
Each drug pair in the SemEval-2013 Task 9 dataset has a corresponding text description. DDIs extraction is a task performed to identify drug pairs in input sentences that describe the interactions of the drug pairs and assign the correct types of interactions to the drug pairs. Herein, BERT [14] was used as the basic model to extract features of drug-drug pairs from texts. The BERT model is very difficult to train, with an astonishing scale of billions of parameters. Researchers typically load pre-trained model parameters before training a specific task. In this study, three pretraining models were compared, and the most suitable model was selected for extracting text-related features. PubMedBERT was eventually integrated into the framework to enhance training efficiency and extract drug pair features more effectively from text. PubMedBERT, a pre-trained model, was trained on the latest collection of PubMed5 abstracts, comprising 21 GB of data with 14 million abstracts and 3.2 billion words. A preprocessed input sentence was converted into a real-valued fixed-size vector via a BERT-based model (Fig 1A). Specifically, given an input sentence S = (w1,…,wn), where drugs d1 and d2 are involved, the sentence was first split into word fragments through the WordPiece algorithm [47] to obtain the corresponding token embedding . Each token embedding
was then converted into a real-valued pretrained contextualized embedding
via the BERT model. In addition, dp-dimensional drug-relative position embeddings
and
were prepared for each word piece, which consists of the relative positions d1 and d2.
and
was concatenated to obtain the corresponding position embeddings for each word fragment:
(10)
where [;] denotes concatenation.
Finally, input sentence text embeddings and input sentence position embeddings
were obtained.
Features fusion methods
In this section, three different feature-fusion methods were used to combine features obtained from subgraphs and texts. Their performances were compared, and the best (Fusion method 2) was selected.
Fusion method 1
A Bi-directional Long Short-Term Memory (BiLSTM) [48] was used to deal with the pre-trained embeddings, and fairly desirable results were obtained (based on Chen et al. [15] and Dou et al. [49] studies). BiLSTM was used to further emphasize contextual features in sentences. ew was input into BiLSTM to obtain forward and reverse sentence representations (Fig 2B) as follows:
(11)
where
,
and
, and dlstm represent the hidden layer size of LSTM.
(B), (C) are candidate methods for Fig 1A. The purple rectangles represent the information from BKG.
The outputs and
were obtained as follows:
(12)
zi was then introduced as a concatenation of k input embeddings around
:
(13)
A convolution operation was performed on zi as follow:
(14)
where ⨀ represents element-wise product: btext represents a bias term; and f(∙) represents a GELU function [50]. A weight tensor for the convolution was defined as
represents the j-th column of Wtext; k represents window size. In addition, max-pooling was used to convert the output of each filter in the convolutional layer into a fixed-size vector as follows:
(15)
Finally, etext was concatenated with hdp as shown in Eq (16):
(16)
where ℋ represents the final result vector, which includes text features and subgraph features about the targeted drug pairs.
Fusion method 2
The basic process for method 2 (Fig 1A) is similar to that of Fusion method 1 (Fig 2B). However, the output of BERT is not passed to BiLSTM, but directly delivered into CNN after concatenating with position embeddings. Therefore, all equations are the same as in Fusion method 1 except for Eq (12), which changes to et = [ew,ep].
Fusion method 3
For fusion method 3 (Fig 2C), the text embeddings ew and drug pairs representations hdp were concatenated as follows:
(17)
Second, and
, representing the positions corresponding to drugs d1 and d2, were concatenated, with the sentence position embedded ep as follows:
(18)
Finally, was obtained to emphasize the drug mention location information for drug pairs representations. e was then sent to CNN as mentioned in Fusion Method 1:
(19)
where ℋ represents the final result vector, including text features and subgraph features about targeted drug pairs.
DDIs extraction using subgraph information
The resulting vector ℋ was used as the input to the prediction layer as follows:
(20)
where s = [s1,…,sm] represent the prediction scores;
represents a weight matrix to convert ℋ into prediction scores, where m represent the number of DDI types. A softmax function was then used to convert s into the probability of possible interactions pt.
Finally, the loss function was calculated based on Eq (1) using pt (defined as Eq (21)).
Experiments
Evaluation metrics.
Two common F1 scores: micro-f1 (F1micro) and macro-f1 (F1macro) score were used to evaluate the performance of the method. These two F1 score metrics can be calculated by precision (P) and recall (R). Micro precision, recall and F1 score were calculated as follows:
(22)
TP, FP and FN represent the number of true positive cases, number of false positive cases, F number of false negative cases, respectively; i represents the i-th DDIs type. Although F1micro is a common metric in DDIs extraction studies, F1micro should give more weight to the primary classification according to Eq (22). In this study, Macro precision, recall, and F1 score were determined as follows:
(23)
F1macro is more at each type, regardless of its sample size according to Eq (23). Opitz [51] (2019) pointed that F1macro is more suitable for unevenly distributed dataset. Moreover, we employ AUC and AUPR metrics for both micro-averaged and macro-averaged evaluations. AUC represents the area under the receiver operating characteristic curve, while AUPR represents the area under the precision-recall curve.
Experimental setting.
PyTorch library is widely used to implement SubGE-DDI. In this study, PubMedBERT was used to convert tokens into 768 dimensional vectors. In addition, drug node embeddings were initialized by xavier uniform with gain of . The batch size was set to 32, and Adam was used as the optimizer. The initial learning rate and the number of fine-tuning epochs were 5−5 and 5, respectively. Other important settings are shown in Table 5.
Results and discussion
Model performance
The SubGE-DDI model was evaluated based on the test dataset of the SemEval-2013 Task 9. Moreover, we compared the performance to that of other state-of-the-art models (Table 6). Both micro and macro metrics were used to evaluate the performance of the proposed model (F1micro and F1macro scores).
To ensure fair and reproducible comparisons, we applied the same pre-processed data to all models, regardless of their original dataset sizes. Compared with other methods, SubGE-DDI had the best F1 scores, precision and AUPR values on both micro-averaged and macro-averaged metrics, obtaining F1micro, Precisionmicro, AUPRmicro, F1macro, Precisionmacro and AUPRmacro of 83.91%, 85.02%, 90.96%, 84.75%, 89.22%, and 73.00%, respectively. The results of SRGU-CNN [52] and RHCNN [37] demonstrated that the models using pre-training method achieved better performance compared to models using other machine learning algorithms. In addition, SubGE-DDI outperformed the models that used BERT only (SciBERT [53], BioBERT, and PubMedBERT), further indicating that the subgraph feature can improve DDIs extraction.
The average F1 score on micro-averaged metrics for five-folds of cross-validated training dataset is shown in Table 7. The model with position features and subgraph features scored higher on all categories than the baseline models. This finding indicates that the proposed framework can effectively alleviate the category imbalance problem.
The results of ROC curve (receiver operating characteristic curve) and PR curve (precision-recall curve) are shown in Fig 3. Although the model had a strong DDIs extraction ability, the category imbalance still existed (Fig 3B).
ROC curve and PR curve on SemEval-2013 Task 9 test set, (A) ROC curve, (B) PR curve.
Ablation experiments
Ablation experiments were performed to explore the role of each component in SubGE-DDI. These experiments included models that use only text features as a baseline, followed by comparison of both micro-averaged and macro-averaged metrics by adding or removing additional features. In this paper, the performance of the model was worst when it was only based on PubMedBERT. Moreover, its performance did not significantly improve following addition of the position embeddings (Table 8). These findings indicate that text features are not significantly related to position embeddings. However, the model showed better performance when the subgraph features were added to the model than when the position features were added. Notably, the best performance was achieved when both position features and subgraph features were added. Compared with the sum of improvement after the addition of position features and subgraph features separately, the improvement was greater when position features and subgraph features were added together (F1micro: 1.24% vs. -0.23% and 0.5%; F1macro: 1.90% vs. 0.08% and 0.69%). These findings indicate that the addition of position features can help subgraph features improve the performance of DDIs extraction models. Interestingly, the model integrating subgraphs achieved the highest recall scores for both micro-averaged and macro-averaged metrics. This indicates that leveraging information from drug pair knowledge subgraphs can improve the capacity of the model to identify potential DDI relationships, hence obtain the highest recall.
In addition, the three new fusion methods were compared to obtain the best method. The three fusion methods were evaluated using both micro-averaged and macro-averaged metrics based on the SemEval-2013 Task 9 test set, as shown in Table 9.
The Fusion Method 2 had the best performance on both micro-averaged and macro-averaged metrics (Table 9). Therefore, Method 2 was selected.
Finally, to obtain the most suitable pre-trained language model for our experiment, we analyzed them on three different pre-trained language models: SciBERT, BioBERT and PubMedBERT. Each model was evaluated using the F1 scores of both micro-averaged and macro-averaged metrics when using different feature combinations as shown in Table 10.
Analysis of the results showed that the three pre-trained language models had better performance when all the features were combined compared with ‘BERT-only’. However, because the PubMedBert model obtained the best F1 scores on both micro-averaged and macro-averaged metrics, it was selected as the pre-trained language model.
The influence of different loss functions
To determine the impact of the multi-focal loss function, we compared the F1 score on binary classification between the model with cross entropy loss and the model with multi-focal loss as indicated in Fig 4.
The results demonstrate that the multi-focal loss has better performance on all DDI types except “Int.” compared to the cross-entropy loss. Moreover, the performance of multi-focal loss on “Int.” is similar to that of the cross-entropy loss (57.81±4.92% vs 57.90±6.05%). This suggests that the multi-focal loss can alleviate category imbalance in multi-classification tasks and improve the Negative, Mechanism, Effect and Advise, by 0.04%, 1.12%, 0.36%, and 3.86%, respectively.
Error analysis
Our model exhibits a consistent trend of higher precision than recall. To investigate the underlying causes of this pattern, we conducted a comprehensive error analysis. This involved calculating the confusion matrix both before and after normalizing the test set, as depicted in Fig 5. Most importantly, from Fig 5B, we notice that many “Int.” samples were incorrectly classified as “Effect”, while the “Effect” samples were rarely misidentified as “Int.”. Analysis of Eq (22) and Eq (23) indicated that there are too few true positive samples for “Int.”. This caused a low recall rate of “Int.”, and thus reduce the overall recall rate. Subsequently, we reviewed the definition of labels in the SemEval-2013 Task 9 dataset, where the “Int.” was described as “This type is used when a DDI appears in the text without providing any additional information”. Therefore, “Int.” is a vague category. If a specific term in the sentence unambiguously refers to a particular DDI category other than "Int.", it may lead to a misclassification. To further confirm this point, two examples of “Int.” are provided here, the first correctly classified and the second misclassified as “Effect”:
- ‘Drug1 may interact with Drug2 or mitotane (causing too great a decrease in adrenal function).’
- ‘Drug1 may interact with aminoglutethimide or Drug2 (causing too great a decrease in adrenal function).’
The initial sentence shows a similarity between drug1 and drug2 as they both have the term "interact" associated with them. Conversely, in the second sentence, the word "decrease" is more proximate to drug2 than the term "interact". This proximity might cause confusion for the model, potentially leading to an incorrect categorization as ’Effect’.
(A): Confusion matrix without Normalization, (B): Confusion matrix with Normalization.
Case study
In addition, we show four cases of prediction results as case study. Case 1, 2 and 3 are predicted correctly by using SubGE-DDI but incorrectly when PubmedBert is used alone (Table 11). Case 4 has opposite results. The result (like case 1, case 2) present a scenario where PubmedBert shows an incorrect prediction and SubGE-DDI exhibit correct prediction although the distance between DRUG1 and DRUG2 is close.
Hence, leveraging information about drug pairs in BKG may be beneficial in cases where predictions relying on their contextual surroundings are difficult. Nevertheless, we infer that information obtained from BKG may result in inaccurate predictions when the term "effects" is in close proximity to DRUG1 in case 4. Despite this, SubGE-DDI still showed good performance in practical tasks.
Conclusion
In this study, a novel framework called SubGE-DDI, which combines the external subgraph features of BKG and semantic text features is proposed for DDIs prediction in biomedical text. Our experimental results indicate that the developed framework can efficiently predict DDIs, indicating that the subgraph features obtained from relevant knowledge graphs can improve DDIs extraction.
SubGE-DDI incorporates external subgraph features from the BKG into biomedical text features thereby achieving superior performance compared to other similar methods. To achieve this, SubGE-DDI employs a novel GCN-based model called SubAGCN to extract relevant information about the drug pairs in sentences from the BKG. During this extraction process, SubGE-DDI utilizes a self-attention module to filter out irrelevant information and retain only the important aspects, resulting in more accurate and effective DDI extraction. Our experiments revealed two limitations of SubGE-DDI. First, its effectiveness is influenced by the quality of extracted subgraphs. Second, the performance of DDI extraction can be hampered to a certain extent by class imbalance.
To address these limitations, we will focus on the following aspects in future. Firstly, we will shall develop more efficient subgraph extraction methods, to improve the quality of subgraphs and hence the performance of SubGE-DDI. Thus, meta-paths or adaptive propagation depth decision strategies will be adopted [55,56]. In addition, we plan to investigate the application of data augmentation techniques to facilitate DDI extraction.
References
- 1.
Baxter Karen, Preston Claire L., editors. Stockley’s drug interactions. Vol. 495. London: Pharmaceutical Press; 2010.
- 2. Agarwal S, Agarwal V, Agarwal M, Singh M. Exosomes: Structure, Biogenesis, Types and Application in Diagnosis and Gene and Drug Delivery. Curr Gene Ther. 2020; 20(3):195–206. pmid:32787759.
- 3. Wishart DS, Feunang YD, Guo AC, Lo EJ, Marcu A, Grant JR, et al. DrugBank 5.0: a major update to the DrugBank database for 2018. Nucleic Acids Res. 2018 Jan 4;46(D1):D1074–D1082. pmid:29126136; PMCID: PMC5753335.
- 4. Tatonetti NP, Fernald GH, Altman RB. A novel signal detection algorithm for identifying hidden drug-drug interactions in adverse event reports. J Am Med Inform Assoc. 2012 Jan-Feb;19(1):79–85. Epub 2011 Jun 14. pmid:21676938; PMCID: PMC3240755.
- 5. Xiong G, Yang Z, Yi J, Wang N, Wang L, Zhu H, et al. DDInter: an online drug-drug interaction database towards improving clinical decision-making and patient safety. Nucleic Acids Res. 2022 Jan 7;50(D1):D1200–D1207. pmid:34634800; PMCID: PMC8728114.
- 6. Kanehisa M, Furumichi M, Sato Y, Kawashima M, Ishiguro-Watanabe M. KEGG for taxonomy-based analysis of pathways and genomes. Nucleic Acids Res. 2023 Jan 6;51(D1):D587–D592. pmid:36300620; PMCID: PMC9825424.
- 7. Zitnik M, Sosic R, Leskovec J. Biosnap datasets: Stanford biomedical network dataset collection. Note: http://snap.stanford. edu/biodata, 2018.
- 8. Hu W, Zhang W, Zhou Y, Luo Y, Sun X, Xu H, et al. MecDDI: Clarified Drug-Drug Interaction Mechanism Facilitating Rational Drug Use and Potential Drug-Drug Interaction Prediction. J Chem Inf Model. 2023 Mar 13;63(5):1626–1636. Epub 2023 Feb 19. pmid:36802582.
- 9. Vilar S, Friedman C, Hripcsak G. Detection of drug-drug interactions through data mining studies using clinical sources, scientific literature and social media. Brief Bioinform. 2018 Sep 28;19(5):863–877. pmid:28334070; PMCID: PMC6454455.
- 10.
Isabel Segura-Bedmar, Paloma Martínez, María Herrero-Zazo. 2013. SemEval-2013 Task 9: Extraction of Drug-Drug Interactions from Biomedical Texts (DDIExtraction 2013). In Second Joint Conference on Lexical and Computational Semantics (*SEM), Volume 2: Proceedings of the Seventh International Workshop on Semantic Evaluation (SemEval 2013), pages 341–350, Atlanta, Georgia, USA. Association for Computational Linguistics.
- 11. Liu S, Tang B, Chen Q, Wang X. Drug-Drug Interaction Extraction via Convolutional Neural Networks. Comput Math Methods Med. 2016;2016:6918381. Epub 2016 Jan 31. pmid:26941831; PMCID: PMC4752975.
- 12. Zheng W, Lin H, Luo L, Zhao Z, Li Z, Zhang Y, et al. An attention-based effective neural model for drug-drug interactions extraction. BMC Bioinformatics. 2017 Oct 10;18(1):445. pmid:29017459; PMCID: PMC5634850.
- 13. Christopoulou F, Tran TT, Sahu SK, Miwa M, Ananiadou S. Adverse drug events and medication relation extraction in electronic health records with ensemble deep learning methods. J Am Med Inform Assoc. 2020 Jan 1;27(1):39–46. pmid:31390003; PMCID: PMC6913215.
- 14.
Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
- 15. Chen J, Sun X, Jin X, Sutcliffe R. Extracting drug-drug interactions from no-blinding texts using key semantic sentences and GHM loss. J Biomed Inform. 2022 Nov;135:104192. Epub 2022 Sep 3. pmid:36064114.
- 16. Lee J, Yoon W, Kim S, Kim D, Kim S, So CH, et al. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics. 2020 Feb 15;36(4):1234–1240. pmid:31501885; PMCID: PMC7703786.
- 17. Molina M, Jiménez C, Montenegro C. Improving Drug-Drug Interaction Extraction with Gaussian Noise. Pharmaceutics. 2023 Jun 26;15(7):1823. pmid:37514010; PMCID: PMC10385013.
- 18.
Ostendorff, M., Bourgonje, P., Berger, M., Moreno-Schneider, J., Rehm, G., Gipp, B. Enriching bert with knowledge graph embeddings for document classification[J]. arXiv preprint arXiv:1909.08402, 2019
- 19.
Wang, S., Pan, Y., Xu, Z., Hu, B., Wang, X. Enriching BERT With Knowledge Graph Embedding For Industry Classification[C]//Neural Information Processing: 28th International Conference, ICONIP 2021, Sanur, Bali, Indonesia, December 8–12, 2021, Proceedings, Part VI 28. Springer International Publishing, 2021: 709–717.
- 20.
Masaki Asada, Makoto Miwa, Yutaka Sasaki. Enhancing drug-drug interaction extraction from texts by molecular structure information[J]. arXiv preprint arXiv:1805.05593, 2018.
- 21. Asada M, Miwa M, Sasaki Y. Using drug descriptions and molecular structures for drug-drug interaction extraction from literature. Bioinformatics. 2021 Jul 19;37(12):1739–1746. pmid:33098410; PMCID: PMC8289381.
- 22. Lin X, Dai L, Zhou Y, Yu ZG, Zhang W, Shi JY, et al. Comprehensive evaluation of deep and graph learning on drug-drug interactions prediction. Brief Bioinform. 2023 Jul 20;24(4):bbad235. pmid:37401373.
- 23. Ren ZH, You ZH, Yu CQ, Li LP, Guan YJ, Guo LX, et al. A biomedical knowledge graph-based method for drug-drug interactions prediction through combining local and global features with deep neural networks. Brief Bioinform. 2022 Sep 20;23(5):bbac363. pmid:36070624.
- 24. Su X, Hu L, You Z, Hu P, Zhao B. Attention-based Knowledge Graph Representation Learning for Predicting Drug-drug Interactions. Brief Bioinform. 2022 May 13;23(3):bbac140. pmid:35453147.
- 25. Gu J, Bang D, Yi J, Lee S, Kim DK, Kim S. A model-agnostic framework to enhance knowledge graph-based drug combination prediction with drug-drug interaction data and supervised contrastive learning. Brief Bioinform. 2023 Sep 20;24(5):bbad285. pmid:37544660.
- 26. Su X., You Z., Huang D., Wang L., Wong L., Ji B., et al. Biomedical Knowledge Graph Embedding With Capsule Network for Multi-Label Drug-Drug Interaction Prediction. IEEE Transactions on Knowledge and Data Engineering, vol. 35, no. 6, pp. 5640–5651, 1 June 2023,
- 27. Tang Z., Chen G., Yang H., Zhong W., C. Y. -C. Chen. DSIL-DDI: A Domain-Invariant Substructure Interaction Learning for Generalizable Drug-Drug Interaction Prediction. IEEE Trans Neural Netw Learn Syst. 2023 Feb 14;PP. Epub ahead of print. pmid:37022856.
- 28. Duan B, Peng J, Zhang Y. IMSE: interaction information attention and molecular structure based drug drug interaction extraction. BMC Bioinformatics. 2022 Aug 14;23(Suppl 7):338. pmid:35965308; PMCID: PMC9375903.
- 29. He H, Chen G, Yu-Chian Chen C. 3DGT-DDI: 3D graph and text based neural network for drug-drug interaction prediction. Brief Bioinform. 2022 May 13;23(3):bbac134. pmid:35511112.
- 30. Tatonetti NP, Ye PP, Daneshjou R, Altman RB. Data-driven prediction of drug effects and interactions. Sci Transl Med. 2012 Mar 14;4(125):125ra31. pmid:22422992; PMCID: PMC3382018.
- 31. Kuhn M, Letunic I, Jensen LJ, Bork P. The SIDER database of drugs and side effects. Nucleic Acids Res. 2016 Jan 4;44(D1):D1075–9. Epub 2015 Oct 19. pmid:26481350; PMCID: PMC4702794.
- 32. Jewison T, Su Y, Disfany FM, Liang Y, Knox C, Maciejewski A, et al. SMPDB 2.0: big improvements to the Small Molecule Pathway Database. Nucleic Acids Res. 2014 Jan;42(Database issue):D478-84. Epub 2013 Nov 6. pmid:24203708; PMCID: PMC3965088.
- 33. Avram S, Bologa CG, Holmes J, Bocci G, Wilson TB, Nguyen DT, et al. DrugCentral 2021 supports drug discovery and repositioning. Nucleic Acids Res. 2021 Jan 8;49(D1):D1160–D1169. pmid:33151287; PMCID: PMC7779058.
- 34. Brown GR, Hem V, Katz KS, Ovetsky M, Wallin C, Ermolaeva O, et al. Gene: a gene-centered information resource at NCBI. Nucleic Acids Res. 2015 Jan;43(Database issue):D36-42. Epub 2014 Oct 29. pmid:25355515; PMCID: PMC4383897.
- 35. Davis AP, Grondin CJ, Johnson RJ, Sciaky D, Wiegers J, Wiegers TC, et al. Comparative Toxicogenomics Database (CTD): update 2021. Nucleic Acids Res. 2021 Jan 8;49(D1):D1138–D1143. pmid:33068428; PMCID: PMC7779006.
- 36. Gu Yu, Tinn Robert, Cheng Hao, Lucas Michael, Usuyama Naoto, Liu Xiaodong, et al. (2022). Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing. ACM Transactions on Computing for Healthcare. 3. 1–23.
- 37. Sun X, Dong K, Ma L, Sutcliffe R, He F, Chen S, et al. Drug-Drug Interaction Extraction via Recurrent Hybrid Convolutional Neural Networks with an Improved Focal Loss. Entropy (Basel). 2019 Jan 8;21(1):37. pmid:33266753; PMCID: PMC7514143.
- 38. Quan C, Hua L, Sun X, Bai W. Multichannel Convolutional Neural Network for Biological Relation Extraction. Biomed Res Int. 2016;2016:1850404. Epub 2016 Dec 7. pmid:28053977; PMCID: PMC5174749.
- 39. Zhu Y, Li L, Lu H, Zhou A, Qin X. Extracting drug-drug interactions from texts with BioBERT and multiple entity-aware attentions. J Biomed Inform. 2020 Jun;106:103451. Epub 2020 May 23. pmid:32454243.
- 40. Lin TY, Goyal P, Girshick R, He K, Dollar P. Focal Loss for Dense Object Detection. IEEE Trans Pattern Anal Mach Intell. 2020 Feb;42(2):318–327. Epub 2018 Jul 23. pmid:30040631.
- 41.
Muhan Zhang, Yixin Chen. 2018. Link prediction based on graph neural networks. In Proceedings of the 32nd International Conference on Neural Information Processing Systems (NIPS’18). Curran Associates Inc., Red Hook, NY, USA, 5171–5181.
- 42.
Marc Brockschmidt. 2020. GNN-FiLM: graph neural networks with feature-wise linear modulation. In Proceedings of the 37th International Conference on Machine Learning (ICML’20), Vol. 119. JMLR.org, Article 107, 1144–1152.
- 43. Yu Y, Huang K, Zhang C, Glass LM, Sun J, Xiao C. SumGNN: multi-typed drug interaction prediction via efficient knowledge graph summarization. Bioinformatics. 2021 Sep 29;37(18):2988–2995. pmid:33769494; PMCID: PMC10060701.
- 44.
Peter Shaw, Jakob Uszkoreit, Ashish Vaswani. 2018. Self-Attention with Relative Position Representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers), pages 464–468, New Orleans, Louisiana. Association for Computational Linguistics.
- 45.
Schlichtkrull M., Kipf T.N., Bloem P., van den Berg R., Titov I., Welling M. (2018). Modeling Relational Data with Graph Convolutional Networks. In:, et al. The Semantic Web. ESWC 2018. Lecture Notes in Computer Science(), vol 10843. Springer, Cham. https://doi.org/10.1007/978-3-319-93417-4_38
- 46.
Xu, K., Li, C., Tian, Y., Sonobe, T., Kawarabayashi, K., Jegelka, S. (2018). Representation Learning on Graphs with Jumping Knowledge Networks. Proceedings of the 35th International Conference on Machine Learning, in Proceedings of Machine Learning Research 80:5453–5462. Available from https://proceedings.mlr.press/v80/xu18c.html.
- 47.
Taku Kudo, John Richardson. 2018. SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 66–71, Brussels, Belgium. Association for Computational Linguistics.
- 48.
Huang Z, Xu W, Yu K. Bidirectional LSTM-CRF models for sequence tagging[J]. arXiv preprint arXiv:1508.01991, 2015.
- 49. Dou M, Ding J, Chen G, Duan J, Guo F, Tang J. IK-DDI: a novel framework based on instance position embedding and key external text for DDI extraction. Brief Bioinform. 2023 May 19;24(3):bbad099. pmid:36932655.
- 50.
Hendrycks D, Gimpel K. (2016) Gaussian error linear units (GELUs). arXiv Preprint arXiv: 1606.08415.
- 51.
Opitz, J., Burst, S. (2019). Macro f1 and macro f1. arXiv preprint arXiv:1911.03347.
- 52. Wu H, Xing Y, Ge W, Liu X, Zou J, Zhou C, et al. Drug-drug interaction extraction via hybrid neural networks on biomedical literature. J Biomed Inform. 2020 Jun;106: 103432. Epub 2020 Apr 23. pmid:32335223.
- 53.
Iz Beltagy, Kyle Lo, Arman Cohan. 2019. SciBERT: A Pretrained Language Model for Scientific Text. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 3615–3620, Hong Kong, China. Association for Computational Linguistics.
- 54.
Xin Jin, Xia Sun, Jiacheng Chen, Richard Sutcliffe. 2022. Extracting Drug-drug Interactions from Biomedical Texts using Knowledge Graph Embeddings and Multi-focal Loss. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management (CIKM ’22). Association for Computing Machinery, New York, NY, USA, 884–893. https://doi.org/10.1145/3511808.3557318
- 55. Zhao B., Wang L., Hu P., Wong L, Su X, Wang B, et al. (2023). Fusing Higher and Lower-order Biological Information for Drug Repositioning via Graph Representation Learning. IEEE Transactions on Emerging Topics in Computing.
- 56. Zhao BW, Su XR, Hu PW, Huang YA, You ZH, Hu L. iGRLDTI: an improved graph representation learning method for predicting drug-target interactions over heterogeneous biological information network. Bioinformatics. 2023 Aug 1;39(8): btad451. pmid:37505483; PMCID: PMC10397422.