Figures
Abstract
In the era of digital communication, the rapid spread of information has brought both benefits and challenges. While it has democratized access to knowledge, it has also led to an increase in fake news, with significant societal repercussions. The COVID-19 pandemic has exacerbated this issue, resulting in what the World Health Organization has termed an “infodemic." In light of this, developing effective methods for detecting fake news is of paramount importance. In this paper, we introduce a novel approach that integrates knowledge graphs and Named Entity Recognition (NER) based on a biomedical language model to address the challenge of fake news detection. Our method aims to enhance detection accuracy by combining content analysis with entity-level insights. Our approach involves three key components. First, content analysis uses a contextual language model to capture the semantic context of the content, enabling the extraction of meaningful insights essential for identifying fake news. Second, the NER component, built on a biomedical language model, precisely identifies and categorizes entities within the content, offering a deeper understanding crucial for detecting misinformation in the biomedical domain. Finally, entity integration employs knowledge graph embeddings to transform identified entities into a format that facilitates enhanced processing and detection. By blending these components, our method creates a unified representation of the content, incorporating both semantic context and entity-based insights. This comprehensive approach significantly improves the accuracy of fake news detection. Our extensive experiments demonstrate the effectiveness of this method, particularly in the early identification of false information. The results underscore the potential of our approach as a powerful tool in combating misinformation, particularly in critical areas such as public health.
Citation: Dar RA, Hashmy R, Anwar MS, Böhm P, Frnda J (2025) Semantic knowledge graph fusion for fake news detection: Unifying content-based features and evidence-based analysis in the COVID-19 infodemic. PLoS One 20(7): e0321919. https://doi.org/10.1371/journal.pone.0321919
Editor: Venkatachalam Kandasamy, University of Hradec Kralove, CZECHIA
Received: October 19, 2023; Accepted: March 13, 2025; Published: July 1, 2025
Copyright: © 2025 Dar et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The datasets and ontologies utilized in this study are publicily available, and have been deposited in a Zenodo repository along with the code for implementation. They are accessible at https://doi.org/10.5281/zenodo.12669980.
Funding: The author(s) received no specific funding for this work.
Competing interests: The authors have declared that no competing interests exist.
Abbreviations: MLP:, Multilayer perceptron; NER:, Named entity recognition; Kg:, Knowledge hraph; COVID-19:, Coronavirus Disease 2019; CORD-19:, COVID-19 Open Research Dataset; SVM:, Support vector machine; TF-idf:, Term frequency-inverse document frequency; LR:, Logistic regression; F1:, F1 score
Introduction
In the ever-evolving landscape of information dissemination, the insidious spread of fake news has become a pervasive concern, particularly on influential social media platforms such as Twitter [1–3]. The transformative impact of these platforms in shaping public opinion amplifies the urgency of addressing the rapid proliferation of false information [4,5]. Beyond the immediate threat of misinformation, fake news engenders far-reaching consequences, including financial, social, and emotional risks to society at large [6,7]. Consider, for instance, the financial implications of fake news influencing stock markets. In recent years, incidents where misinformation on social media led to sudden stock fluctuations have underscored the tangible economic consequences [8]. Furthermore, the societal impact of politically motivated fake news, capable of inciting unrest or influencing election outcomes, is a critical concern. In addition to the financial and societal impacts, the dissemination of fake news undermines public trust in institutions and erodes social cohesion [9]. During times of crisis, such as public health emergencies or political upheavals, misinformation can sow confusion and division among communities. When individuals are exposed to false or misleading information, it distorts their understanding of reality and fosters skepticism toward authoritative sources of information. Consequently, public trust in institutions such as governments, media organizations, and scientific establishments is undermined, impeding efforts to address pressing issues and hindering societal progress [10]. Therefore, combating fake news is essential for preserving the integrity of information and fostering trust and solidarity within society. Navigating the intricate landscape of fake news detection demands innovative methodologies, especially in the face of increasingly sophisticated misinformation techniques. Traditional approaches, often centered around language analysis or predefined patterns, prove inadequate against the intricacies of well-crafted manipulation strategies. For instance, analyzing linguistic patterns alone might fall short in identifying subtly crafted fake news pieces designed to mimic legitimate reporting [11]. Hence, there is a critical need for innovative and versatile detection frameworks capable of adapting to evolving patterns of deception. By leveraging advanced technologies such as natural language processing, machine learning, and knowledge representation, researchers can develop sophisticated algorithms that identify fake news and anticipate emerging trends in misinformation dissemination. Moreover, the incorporation of domain-specific knowledge is paramount. Consider the domain of health-related misinformation during a pandemic, where erroneous information about preventive measures, treatments, or vaccine efficacy can have dire consequences. In response, this study advocates for a comprehensive strategy that integrates Knowledge Graphs (KG) and harnesses—BioBERT [12]—based Named Entity Recognition (NER) in tandem with linguistic analysis to enhance detection accuracy. Drawing inspiration from recent advancements, such as the utilization of KGs in enhancing contextual understanding, we suggest that incorporating background knowledge is pivotal in the fight against misinformation [13]. To illustrate the multifaceted nature of misinformation, consider the scenario of health-related fake news. During a public health crisis, misinformation regarding treatment methods or vaccine efficacy can lead to real-world consequences, such as individuals opting for unverified treatments or refusing vaccination.
The proposed methodology, encapsulated in the CogiGraph framework, comprises three fundamental components: (1) Content Encoding, (2) Named Entity Recognition, and (3) Knowledge Graph Integration. These components synergistically contribute to a dual objective—capturing nuanced meanings within news content and comprehending intricate relationships among entities, empowering accurate decisions in fake news detection. Delving into the Content-Encoding component, we employ DistilBERT [14] to uncover semantic nuances embedded in news articles. This approach allows us to analyze the contextual subtleties often exploited in fake news to mimic genuine reporting [15]. The Named Entity Recognition (NER) component utilizes a BioBERT-based model, allowing for precisely identifying entities within the text. This is crucial in detecting misinformation that might manipulate the representation of entities or events [16]. Finally, the third component utilizes simplE [17] for graph embeddings, facilitating the seamless integration of Knowledge Graphs to enhance the understanding of relationships among entities and topics. This becomes particularly relevant in scenarios where fake news leverages complex networks of entities to propagate misinformation, as seen in politically motivated campaigns [18]. To evaluate the efficacy of the CogiGraph framework, comprehensive analyses are conducted on the “Constraint@AAAI 2021 COVID-19” dataset [19]. The results showcase the superior performance of our framework compared to state-of-the-art methods. In an era where the urgency for early detection of fake news intensifies, this paper makes a significant contribution by introducing a comprehensive framework. By seamlessly combining Knowledge Graphs and BioBERT-based Named Entity Recognition (NER) with detailed content analysis, our approach focuses solely on textual content to enable early and accurate detection of fake news.
In conclusion, as the rapid dissemination of misinformation challenges the credibility of information, this research endeavors to lay the groundwork for fostering credible and factual information. By integrating domain-specific knowledge graphs curated from reputable ontology and datasets, our detection framework gains the ability to discern the validity of health-related claims with higher accuracy. Additionally, utilizing specialized named entity recognition models fine-tuned on domain-specific corpora further enhances the precision of entity identification within health-related news articles. Therefore, our approach not only addresses the broader challenge of fake news detection but also offers tailored solutions to combat misinformation within specific domains, thereby safeguarding public health and well-being to navigate the evolving information landscape contributing to the ongoing dialogue on fortifying the integrity of our digital discourse.
Related work
Researchers have proposed various detection methods to address the challenge of fake news. This section reviews existing literature that analyzes content, employs knowledge-based methods, and integrates contextual information using knowledge graphs.
Content-based approaches
Content analysis methods, focusing on the language and structure of news, have gained prominence in fake news detection [20]. Some studies utilized conventional machine learning algorithms for misinformation detection. For instance, hybrid features and decision trees achieved high accuracy in Twitter propaganda detection [21], highlighting the potential of content-based features. Advanced deep learning techniques, such as attention mechanisms, delve into the intricate details of language and context [22]. For example, a study investigates the potential of Long Short-Term Memory (LSTM) networks for content-based spam detection on social media. Another hybrid deep learning model, combining convolutional and recurrent neural networks (RNNs), demonstrated superior performance in fake news classification [23].
Recent research has explored the impact of various embedding techniques on fake news detection. The use of word embeddings has been shown to enhance detection accuracy by capturing semantic relationships between words [24]. Transformer-based embeddings, such as those derived from models like BERT, have also been applied to fake news detection tasks, leveraging their ability to model complex language patterns [25]. Furthermore, document embeddings, which represent entire texts as vectors, have been utilized to improve detection performance by encapsulating contextual information [26].
In addition to embedding techniques, real-time architectures have been proposed to address the need for timely fake news detection. These systems are designed to process and analyze news content rapidly, enabling prompt identification of misinformation [27]. Moreover, incorporating social network features, such as user interactions and propagation patterns, has been found to enhance detection models by providing additional context beyond the content itself [28].
The challenge of detecting fake news in multilingual contexts has also been addressed through the use of multilingual transformers. These models are capable of understanding and processing multiple languages, making them suitable for detecting misinformation across diverse linguistic landscapes [29].
While these approaches show promise, concerns about bias and limited use of external knowledge persist. Bias from potentially biased training data and fact-tempering attacks can lead to inaccurate predictions and hinder generalizability [30,31]. Potthast et al. [32] identified false information by analyzing the distinct characteristics of textual material using a meta-learning methodology. Deep neural networks have also been used to acquire detection features, circumventing the expensive process of manually designing features. Kong et al. [33] used a combination of bidirectional long short-term memory (Bi-LSTM) and a convolutional neural network (CNN) to identify bogus news accurately. This approach was chosen due to the capacity of these models to represent textual material efficiently. Zhao et al. [34] used a mixture-of-experts model to integrate detection characteristics from many domains, improving performance.
Knowledge-based approaches
Knowledge-based approaches use external information sources and knowledge databases to verify the factual accuracy of claims. Early research employed web-based statistical analysis for fact-checking [35]. Computational fact-checking introduced automatic extraction and verification of claims against structured knowledge databases [36]. Knowledge graphs have been leveraged to navigate relationships between entities and verify factual accuracy [37]. Hybrid systems analyzing linguistic features and external knowledge sources have shown improved fake news detection performance [38]. While knowledge-based approaches demonstrate effectiveness, challenges remain, including ensuring data reliability, addressing limited knowledge base coverage, and adapting to the evolving nature of fake news. Future research directions involve developing robust methods for assessing data reliability, expanding knowledge base coverage, and investigating adaptable machine learning models [39]. Zhang et al. [40] introduced a multimodal knowledge-aware event memory network to identify rumors. More precisely, a knowledge-aware network was built to incorporate external information from real-world knowledge graphs as supplementary evidence. Furthermore, they developed an event memory network to acquire event-invariant characteristics as a benchmark for achieving more resilient representations. A system was developed in [41] that used a knowledge graph to identify and explain bogus news. The retrieved graph embeddings were merged with a graph convolutional network to get the detection outcomes. Li et al. [42] used factual information and subjective opinions to identify false news by creating diverse graph structures. While including external knowledge into the detection process might enhance the reliability of findings by examining the links between items in the knowledge graph, the specific approach to combining text information with external knowledge remains unresolved.
Network immunization for fake news detection
Network immunization techniques aim to prevent the spread of misinformation in social networks by identifying and mitigating fake news at the source. These methods leverage graph-based strategies to identify influential nodes and control the dissemination of malinformation.
Community detection techniques, for instance, have been used to isolate clusters of users involved in spreading fake news. These approaches analyze network structures to identify communities and immunize them effectively, thereby curbing the propagation of misinformation [43].
Weighted directed spanning trees offer another effective solution for mitigating fake news in real time. By organizing network nodes into hierarchical structures, these trees enable quick identification of fake news sources and facilitate timely intervention [44].
Budget-based immunization algorithms focus on optimizing resource allocation to achieve maximum impact in preventing the spread of fake news. These algorithms prioritize key nodes for immunization based on their influence in the network, ensuring efficient use of limited resources [45].
These strategies demonstrate the potential of network-based methods in complementing content-based and knowledge graph approaches, offering a holistic solution to the fake news detection problem.
Knowledge graphs and fake news detection
Modern approaches leverage graph structures for fake news detection. A Graph-based Markov Chain approach segregates real and fake news articles, utilizing random walks to assess similarity [46]. Knowledge graphs (KGs), interconnected databases containing entities and relationships, have been used in recent studies [47]. Models like TransE [48] and DistMult [49] have been proposed for embedding entities and relationships within KGs. Researchers have explored using KGs for link prediction within the context of fake news detection. KGs have been employed for content-based fake news detection, demonstrating the effectiveness of external knowledge sources [50]. Graph neural networks (GNNs) alongside KGs have been proposed, achieving promising results in fake news detection tasks [51,52]. To identify false news, Vaibhav et al. [53] used document sentences as graph nodes to represent documents as graph structures. They next employed graph attention networks to acquire document characteristics. In addition, news shared on social media platforms may include many types of information, such as text, user, and temporal data, that may be used for detecting purposes. Nguyen et al. [54] introduced a technique for learning graphical representations that accurately capture the social context of false news. Zhang et al. [55] developed a heterogeneous network that incorporates news items, authorship, and news topics. They established a deep, diffusive network to combine this information to detect false news. Furthermore, false information often disseminates rapidly via social media platforms, text messaging, or electronic mail. Hence, by using deep learning methods, the identification of false news may be achieved by examining the velocity and extent of news transmission. Additional studies, such as [56], have suggested using graph convolutional networks to represent the spread of news. Dou et al. [57] established a false news detection system that considers user preferences and integrates the spread of news and associated topics. Despite advancements, challenges remain in effectively considering the complex relationships within news articles and interconnected KG information. Future research involves developing robust methods for integrating diverse types of knowledge from KGs, exploring advanced machine learning models, and designing efficient algorithms for large-scale KGs and real-time fake news detection systems. Our approach builds upon these foundations by integrating content-based analysis, knowledge-based validation, and KG-driven enrichment. By blending entities and relationships from KGs with content-based representations, our method aims for a comprehensive understanding of news articles, enhancing the effectiveness of fake news detection, particularly in early detection scenarios.
Methodology
In this section of our research paper, the authors explain the approach we’ve developed. We outline our steps, demonstrating how we’ve combined different techniques to create a solid foundation for our system.
We first discuss the semantic encoding of the news content, the KG construction and enrichment, the graph embedding extraction, and finally, the fusion and prediction sections.
Text preprocessing
We have used tweet-preprocessor (pypi.org/project/tweet-preprocessor/) library to filter out unnecessary data, such as URLs, emoticons, username handles, etc., from the tweets.
Semantic encoding through DistilBERT
Our methodology utilizes DistilBERT as the primary tool for extracting contextual information from the text. DistilBERT is a more resource-efficient variant of BERT, designed to encode the core content of the text into contextual embeddings. These embeddings serve as mathematical representations of the essential information contained within the text, facilitating further analysis. The process involves tokenizing the input text using DistilBERT’s tokenizer, which breaks the text into a sequence of tokens. This token sequence is then passed through the DistilBERT model, which is composed of stacked bidirectional transformer encoders. These layers enable DistilBERT to capture contextual and bidirectional relationships between words, resulting in rich semantic embeddings. Mathematically, the process can be represented as follows:
where
- X represents the input text.
denotes the contextual embeddings generated by DistilBERT.
This approach allows us to efficiently capture the contextual information within the text, providing a solid foundation for subsequent analysis and classification tasks in our fake news detection system.
Constructing a knowledge graph
Our approach to constructing a Knowledge Graph (KG) for COVID-19 knowledge begins with the selection of the COVID-19 ontology [58] as the foundational framework. This ontology, developed through collaborative efforts across multiple disciplines, is a structured system for organizing essential entities and concepts relevant to COVID-19 research. Formally, the COVID-19 ontology is denoted as .
Initial approach: The construction of the KG follows these steps:
- COVID-19 ontology selection: Choose the COVID-19 ontology
as the basis for KG construction.
- Entity identification and mapping: Identify entities and concepts from
and map them to nodes in the KG schema.
- Relationship establishment: Define relationships within
and map them to edges in the KG.
- Node and edge creation: Create nodes and edges in the KG based on the identified entities, concepts, and relationships.
Refinement and enrichment: Further refinement and enrichment of the KG involve:
- Data integration: Incorporate additional information from scientific literature, public health reports, and other relevant sources to enrich the KG.
- Semantic enhancement: Refine entity types, relationships, and attributes to improve semantic representation within the KG.
- Evaluation and feedback: Iteratively refine the KG based on evaluation metrics and feedback from domain experts to ensure alignment with fundamental COVID-19 facts.
The construction process is formalized as follows:
where denotes the KG construction algorithm utilizing the COVID-19 ontology
. Through this iterative construction, refinement, and enrichment process, we aim to develop a comprehensive and accurate representation of COVID-19 knowledge within the KG framework.
KG enrichment
To enhance the knowledge encoded in our Knowledge Graph (KG), we leveraged the CORD-19 dataset [59], which contains a vast collection of scholarly articles related to COVID-19. We focused on the abstracts and introductions of a subset of papers from BioRxiv and PMC. We employed advanced natural language processing (NLP) techniques for entity extraction and relationship inference. We utilized the BioBERT model [12], a domain-specific language model pre-trained on biomedical text for entity extraction. To adapt BioBERT for our specific task of entity recognition, we fine-tuned it using a processed version of the CORD-19 dataset Processed Version of CORD-19 Dataset for NER: https://www.kaggle.com/datasets/sushilkumarinfo/cord19processeddataset. Entity recognition involves assigning entity labels to input tokens X, achieved through token-level classification facilitated by a linear transformation
and softmax activation, as depicted in Equation 2.
Next, we extracted relationships embedded within textual knowledge using the Stanford OpenIE NLP tool [60], in conjunction with the named entities recognized by the fine-tuned BioBERT model. This process yielded triplets (s,p,o) representing complex relationships within text.
We also exclusively incorporate factual news from the training data to uphold the KG’s integrity and reliability. This deliberate selection ensures the KG’s utility and credibility, particularly when verifying news items. Conversely, fake news from the training dataset is deliberately omitted to maintain the accuracy and trustworthiness of the KG, safeguarding its quality for subsequent analyses. The resulting enriched KG, denoted as , provides a comprehensive depiction of COVID-19 entities and their intricate relationships. A high-level view of our Knowledge Graph is illustrated in Fig 1.
Alignment scores and KG embeddings
Alignment scores and KG embeddings play a pivotal role in our fact verification approach, facilitating the assessment of alignment between news items and factual knowledge stored within the knowledge graph. This evaluation offers insights into the degree of similarity between a news item and the existing knowledge, effectively quantifying its truthfulness concerning the KG. Consider a news item represented by a set of named entities and relationships
, extracted using the finetuned BioBERT model and the OpenIE tool, as detailed in the previous subsection. Alignment scores are computed to gauge the similarity strength of each entity mi compared to entities in the knowledge graph.
The alignment score for each entity mention mi and its corresponding KG mention is calculated as follows:
Here, represents the similarity between the news item mention and the corresponding KG mention, while
denotes the maximum possible length of either mention. We employ a similarity calculation procedure to quantify each entity’s similarity and the relationship between the news items and those in the knowledge graph. This involves obtaining textual representations of the named entities extracted from the news items and the knowledge graph. Subsequently, we compute word embeddings for each named entity and relationship recognized within a news item and for each entity and relationship within the knowledge graph.
Let represent the cosine similarity score between a named entity mi from the news items and its counterpart in the knowledge graph, denoted as
. This similarity score is determined based on the semantic similarity between the textual representations of the entities, computed using Word2Vec embeddings. Mathematically, the cosine similarity between two word vectors v and w is given by:
where represents the dot product of the two vectors, and
and
represent their respective Euclidean norms. To identify the aligned KG entity for each entity mentioned mi, we select the one with the highest alignment score, calculated using Equation 6:
Similarly, alignment scores are computed for relationships rj and matched KG relationships are determined. This process generates sets of aligned entities and aligned relationships
. These aligned entities and relationships collectively contribute to creating a more meaningful representation of the news item, capturing the interconnections between entities. The embedding of the news item based on alignment, denoted as
, is computed by aggregating the embeddings of aligned entities and relationships, with the alignment scores for more accurate contextualization:
Equation 7 represents the combination operation applied to embeddings, where denotes the combined embedding. Here,
and
represent the sizes of the sets of aligned entities and relationships, respectively. The embeddings
and
are extracted using SimplE, a model designed for capturing complex structural information in knowledge graphs. The addition symbol “+” signifies the fusion of different embeddings, achieved through a weighted sum approach. By incorporating alignment scores, aligned entities, and relationships, our fact verification system leverages factual alignment to enhance accuracy and contextual understanding.
The integration of content and knowledge
The next crucial stage in our methodology is concatenating knowledge-based graph embeddings with the contextual embeddings obtained from DistilBERT. This combination is essential for enhancing the representation of claim embeddings. Mathematically, we represent this merging as:
Equation 8 signifies the integration of two essential components: the graph-based embedding capturing alignment-driven context () and the contextual embedding derived from DistilBERT’s semantic understanding (
). The combination of
and
results in a more comprehensive representation of the news item.
provides structured knowledge extracted from the knowledge graph, including relationships and entities relevant to the topic, while
captures the nuanced semantic understanding derived from the textual content of the news item. Integrating these two embeddings allows us to leverage the factual knowledge encoded in the knowledge graph and the contextual understanding derived from the news item’s language. This super representation enables a richer interpretation of the news item, incorporating factual knowledge and contextual relevance.
Predictive modeling
The last step of our method involves using a Multilayer Perceptron (MLP) for final prediction. MLP works well for this task because it combines different types of information. This includes insights from both the content and the graph. In mathematical terms, we can express the MLP’s prediction like this:
In Equation 9, predicts what’s true for the i-th news piece, and
represents the combined information we got from both the content and the graph.
We employed the MLP to classify the news items. It effectively learns from the combined information encapsulated within , allowing it to make informed predictions about the content’s characteristics. Fig 2 offers a visual overview of our methodology. Algorithm 1 outlines various steps for the proposed cogiGraph model.
Algorithm 1. CogiGraph
In subsequent sections, we will delve deeper into the dataset’s characteristics and various implementation details of our methodology, including how the model is evaluated.
Experimental setup
In this section, we discuss how we conducted our experiments to assess the effectiveness of the proposed CogiGraph model for fake news detection. We provide details about the dataset we used, explain how our model is set up, describe how we fine-tuned its parameters, and share the metrics we used to evaluate its performance.
Dataset details
First, we provide insights into the dataset utilized for evaluating our approach: the “Constraint@AAAI 2021 COVID-19” [19] fake news detection dataset. This dataset consists of 10,700 instances, including social media posts and news articles, each labeled as genuine or false. To ensure fair evaluation, the dataset has been divided into training, validation, and test sets in a balanced manner, maintaining similar proportions of genuine and false instances.
The dataset is split into train (60%), validation (20%), test (20%). The dataset is class-wise balanced as 52.34% of the samples consist of real news and 47.66% of the data consists of fake news. Moreover, we maintain the class-wise distribution across train, validation, and test splits.
Genuine news instances were sourced from various articles from reputable news outlets. False news instances were carefully curated using third-party fact-checking sources such as NewsChecker and PolitiFact. This comprehensive dataset allows us to examine the robustness of our approach across various scenarios.
Implementation details
Moving on to the implementation, our framework, the CogiGraph model, is designed to assess the credibility of a news item based solely on the news text and consists of two interconnected modules. The first module employs DistilBERT to generate rich embeddings from news tokens, capturing contextual meaning. The second module leverages a knowledge graph-based approach to represent entities and their relations, enhancing content understanding. These modules collectively form a combined representation that captures content and entity semantics. In this section, we provide a technical overview of the CogiGraph framework.
Data Splitting and Balance: The dataset has been divided into training (60%) , validation(20%) and testing (20%) subsets. In this split, a balanced distribution of genuine and false instances has been maintained across all sets to ensure unbiased evaluation.
Hyperparameter configuration: We used distinct hyperparameters for various components, including DistilBERT, Bio-BERT, Knowledge Graph Enrichment, and MLP Parameters. These hyperparameters were selected based on experimentation to balance computational efficiency and convergence speed.
- DistilBERT: For semantic encoding using DistilBERT, we employed a batch size of 64 and a learning rate of
, determined through experimentation to balance computational efficiency and convergence speed. The model was optimized for a sequence length of 128 tokens, and the fine-tuning process consisted of training for five epochs.
- BioBERT: Within the implementation framework, the Bio-BERT-based Named Entity Recognition (NER) model plays an important role in accurately identifying named entities within news text. The model utilizes the BioBERT model, loaded with pre-trained weights and fine-tuned on a pre-processed version of the CORD-19 dataset specifically curated for recognizing named entities within the COVID-19 domain. We use a dropout layer for regularization and a linear output layer with softmax activation to predict entity labels. Regarding training parameters, we use an AdamW optimizer with a batch size of 64 and a learning rate of
, trained for 30 epochs. The model’s training employs the cross-entropy loss function, optimizing the model through back-propagation while preventing gradient explosion using gradient clipping. Integration of the BioBERT NER model within our framework enhances entity identification.
- Knowledge graph enrichment: Utilizing the fine-tuned BioBERT model for entity recognition and the factual knowledge from a cleaned subset of the CORD-19 dataset consisting of abstracts and introductions of recent scientific research resulted in a rich and structured knowledge graph.
- MLP parameters: The Multilayer Perceptron (MLP), our classification head, comprises two hidden layers comprising 128 units each and utilizing a ReLU activation function. The final classification layer has a softmax activation function for binary classification. We opted for a learning rate of 0.0001 for effective optimization.
Fig 3 demonstrates the effectiveness of the proposed model, as the graph precisely illustrates the performance of both training and validation accuracy. Significantly, a plateau is attained after 150 epochs, highlighting the model’s stability and resilience. During the experiment, a careful incorporation of K-fold cross-validation was used, with a predetermined value of K set at 10, to guarantee a thorough assessment of the model’s performance. Fig 4 visually represents the complex interplay between training and validation loss across several epochs, offering a detailed comprehension of the model’s learning trajectory. The distinguishing factor of the suggested technique lies in its notable attainment of a 98.97% accuracy, accompanied by a strong sensitivity of 99.01% and a formidable F1 score of 98.92. The measurements demonstrate the higher performance of the proposed model, surpassing other state-of-the-art methodologies. tab:1 is a comprehensive compilation of findings, thoroughly juxtaposing the proposed technique and many other innovative methods. The proposed approach was thoroughly assessed using several metrics, such as precision, recall, and the F1 score. This evaluation proved its efficacy and dependability more than current methodologies.
Model training and fusion: Our approach’s effectiveness mostly relies on finetuning, extracting named entities, populating the knowledge graph and the training process, especially given the size of the models and the size of the CORD-19 dataset. To address this, we harnessed the computational power of the NVIDIA AI Server (DGX A100). We used model parallelism, which involves partitioning a neural network model across multiple GPUs for optimized distributed training. Model parallelism allows us to distribute different sub-networks of the model across the available GPUs, thereby overcoming memory limitations and splitting the BioBERT model across multiple available GPUs as its size surpasses the individual memory available to a single GPU.
- Semantic encoding and named entity recognition: Through DistilBERT, we extracted contextual embeddings
from the news content, as shown in equation 1. The BioBERT-based NER model, fine-tuned with the selected hyperparameters, accurately identified COVID-19-relevant named entities. The graph embeddings for the news item X are calculated using equation 7.
- Alignment and fusion: The calculated alignment score and the aligned entities and relationships help us align the claim entities with those of KG entities to get the most proximate ones. The final graph embedding is calculated using the alignment scores and the embeddings of the aligned entities and relationships from the KG. The alignment score and the aligned entities calculation are shown in equations 4 and 6, respectively. The fusion step combines
and KG-based embeddings
into a unified representation E(X), as shown in equation 8.
- Predictive modeling: Our MLP classification head, trained with the chosen hyperparameters, leverages E(X) to predict the authenticity of news items. Using the Adam optimizer with the designated learning rate, combined with the mixture of content and knowledge-driven insights, enables accurate predictions.
The hyperparameters at various stages were selected experimentally to increase the efficiency of the CogiGraph framework. The use of linguistic analysis, together with entity recognition and the knowledge graph, helped us construct a robust system to verify news items from just the news text.
Results and discussion
In this section, we present the results of our experiments and provide an in-depth analysis of the performance of the CogiGraph model. We also compare its performance against other state-of-the-art approaches to assess its efficacy.
Performance comparison
For performance comparison, selecting appropriate metrics is crucial. The metrics are chosen based on their ability to capture various aspects of classification performance, such as accuracy, precision, recall, and F1-score. These metrics ensure a comprehensive evaluation of the model’s ability to distinguish between fake and legitimate news.
For instance, accuracy measures the overall correctness of predictions but may not reflect the performance in imbalanced datasets. Precision and recall, on the other hand, are particularly important for understanding the model’s effectiveness in identifying fake news without false alarms. The F1-score provides a harmonic mean of precision and recall, offering a balanced view when trade-offs exist between these metrics.
In addition, recent studies emphasize the importance of advanced evaluation techniques for fake news detection. For example, one work highlights the need for robust metric selection tailored to specific challenges in misinformation detection, including imbalanced data distributions and varying contexts [72]. Moreover, it is critical to ensure that the metrics align with the objectives of the detection system, whether the focus is on minimizing false negatives (critical in sensitive scenarios) or maximizing true positives (important for broad misinformation campaigns).
Table 1 summarizes the performance of the CogiGraph model and selected state-of-the-art approaches on the “Constraint@AAAI 2021 COVID-19” dataset. The metrics used for evaluation include accuracy, precision, recall, and F1-score because of the balanced nature of the dataset and for better comparison insights.
- Karnyoto et al. [61]: This model comprises a Bidirectional GRU layer receiving input from BERT, followed by an Attention Layer for feature extraction, a Capsule Network Layer for nuanced neuron connections, and a BiGRU-CRF layer for segmentation and processing.
- Sharif et al. [62]: A combination of CNN and BiLSTM was used here. In the CNN component, 64 convolution filters with a kernel size 5x5 are applied, along with a pooling window size of 1x5. The BiLSTM network consists of 32 bidirectional cells with a dropout rate 0.2. These two networks are then sequentially integrated into the combined model.
- Karnyoto et al. [63]: This model incorporates word co-occurrence and TF-IDF to establish edges in three graph models: Graph Convolutional Network, Graph Attention Network, and GraphSAGE. Augmentation techniques like random deletion, insertion, swap, and synonym replacement were applied
- Patwa et al. [19]: This model highlights SVM-based classification as the top-performing method among various machine learning models.
- Sharif et al. [62]: This model leverages a combination of SVM and TF-IDF. The SVM component utilizes TF-IDF features for classification, achieving notable performance.
- Biradar et al. [64]: This model combines a language model with conventional machine learning algorithms using a voting classifier approach, achieving the highest accuracy with an ensemble setting of LR, ULMFit classifier, and BERT classifier.
- Alghamdi et al. [65]: This model leveraged Bidirectional Gated Recurrent Units(BiGRU) over pre-trained CT-BERT architecture.
- Li et al. [66]: The authors utilized a transformer-based architecture and incorporated five-fold five-model cross-validation and the pseudo label algorithm. The findings suggest that the ensemble approach and integrating cross-validation and pseudo-labeling strategies contribute to enhanced performance metrics.
- Raha et al. [67]: The authors went for RoBERTa-base: 12-layer, 768-hidden, 12-heads, 125M parameters. Trained on a larger dataset with increased iterations and a batch size of 8k, RoBERTa removes the NSP objective during pretraining.
- Patwa et al. [68]: The authors have proposed a simple but effective approach to COVID-19 fake news detection, utilizing CT-BERT and ensemble learning techniques. Their experiments validate the effectiveness of BERT-based models in subject-specific tasks, achieving high-quality binary classification.
- Varshney et al. [70]: The authors Employed an Ensemble-based model incorporating Logistic Regression (LR), Linear Support Vector Machine (LSVM), and Classification and Regression Trees (CART), where their voting gives the final decision.
- Das et al. [71]: This method combines metadata with an ensemble of pre-trained language models for fake news classification. It includes Text Preprocessing, Tokenization, Backbone Model Architectures, Ensemble, Statistical Feature Fusion Network, Predictive Uncertainty Estimation Model, and Heuristic Post-Processing.
Our approach outperformed the competition’s leaderboard regarding all the evaluation metrics for this shared task, where this dataset was released initially. Compared with the Statistical Feature Fusion Network with MCDropout (SFFN) and Post-Processing approach [71], our approach performed better in accuracy and recall and achieved a similar F1 score. SFFN is a statistical feature fusion network that incorporates the dropout technique known as Monte Carlo Dropout (MCDropout). This method helps to mitigate overfitting and improve the model’s generalization performance. In the revised manuscript, we will provide a detailed explanation of SFFN and its role in our proposed approach to ensure clarity for readers.
Conclusion
The proposed CogiGraph framework successfully addresses the challenge of fake news detection by integrating content-based features with knowledge graph insights. By leveraging advanced NLP models such as DistilBERT and BioBERT, and enhancing representation through SimplE embeddings, the framework offers a holistic solution to misinformation detection in the biomedical domain. Experimental results demonstrate that CogiGraph outperforms state-of-the-art methods across multiple evaluation metrics, achieving high accuracy and F1 scores.
This research makes several notable contributions:
- Unified framework: The seamless fusion of semantic content analysis and entity-level evidence-based reasoning sets a new benchmark for misinformation detection systems.
- Domain-Specific Effectiveness: By adopting domain-specific biomedical models and knowledge graphs, the system achieves higher accuracy in health-related misinformation scenarios.
- Scalability and Generalizability: The framework is adaptable to various domains beyond COVID-19, providing a strong foundation for future misinformation detection systems.
Limitations
Despite its promising outcomes, the CogiGraph framework has certain limitations:
- Multimodal content handling: The current system primarily focuses on textual data and lacks the capability to process multimedia content, such as images and videos, which are increasingly prevalent in misinformation.
- Computational complexity: Maintaining and updating large-scale knowledge graphs can be computationally intensive and may require significant resources for real-time processing.
- Dynamic knowledge integration: The static nature of the current knowledge graph structure limits its adaptability to real-time updates and evolving information landscapes.
- Domain dependency: The reliance on biomedical-specific models and datasets may reduce the system’s performance in other domains without additional fine-tuning and data preparation.
- Dataset generalizability: Our experiments have been conducted on a single dataset, which may limit the generalizability of the findings. Future work should incorporate additional datasets to further validate the framework across diverse scenarios.
- Explainability: The current framework lacks sophisticated methods for providing transparent and interpretable detection reasoning, which is crucial for practical adoption.
In conclusion, the CogiGraph framework represents a significant advancement in combating misinformation by unifying content and conducting evidence-based analyses. As digital communication continues to evolve, this research lays a foundation for developing sophisticated tools that safeguard information integrity and promote informed public discourse.
Future scope
In the future, CogiGraph holds great promise for tackling fake news. We envision continuously updating its knowledge base with real-time information and incorporating other modalities like images and videos for a richer understanding. To make its reasoning process transparent, we plan to develop methods for explaining predictions and highlighting key information. Additionally, adapting CogiGraph to diverse domains and evaluating it on larger, multi-lingual datasets will enhance its generalizability and fairness. By exploring these avenues, we aim to refine CogiGraph and develop robust tools for combating fake news across various domains and languages, empowering individuals to discern truth from fiction in our information-rich world.
References
- 1. Zhou X, Zafarani R. A survey of fake news. ACM Comput Surv. 2020;53:1–40.
- 2. Kai S, Sliva A, Wang S, Tang J, Liu H. Fake news detection on social media: a data mining perspective. ACM SIGKDD Explor Newslett. 2017;19.
- 3.
Rubin V, Conroy N, Chen Y, Cornwell S. Fake news or truth? Using satirical cues to detect potentially misleading news. In: Proceedings of the Second Workshop on Computational Approaches to Deception Detection. Association for Computational Linguistics; 2016, pp. 7–17.
- 4.
Boczkowski P, Mitchelstein E, Matassi M. Incidental news: how young people consume news on social media. 2017. https://doi.org/10.24251/hicss.2017.217.
- 5.
Flintham M, Karner C, Bachour K, Creswick H, Gupta N, Moran S. Falling for fake news: investigating the consumption of news via social media. In: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. Association for Computing Machinery; 2018, pp. 1–10.
- 6. Grinberg N, Joseph K, Friedland L, Swire-Thompson B, Lazer D. Fake news on Twitter during the 2016 U.S. presidential election. Science. 2019;363:374–8.
- 7. Lazer DMJ, Baum MA, Benkler Y, Berinsky AJ, Greenhill KM, Menczer F, et al. The science of fake news. Science. 2018;359:1094–6.
- 8. Mantere M. Stock market manipulation using cyberattacks together with misinformation disseminated through social media. In: 2013 International Conference on Social Computing, pp. 950–4, 2013.
- 9. Humprecht E. The role of trust and attitudes toward democracy in the dissemination of disinformation—–a comparative analysis of six democracies. Digit J. 2023;:1-–18.
- 10. Montesi M. Understanding fake news during the COVID-19 health crisis from the perspective of information behaviour: the case of Spain. J Librariansh Inf Sci. 2021;53:454–65.
- 11. Mohammad M, Algaraady J, Alrahaili M. Linguistic-based detection of fake news in social media. Int J English Linguistics. 2020;11(1):99.
- 12. Jinhyuk L, Yoon W, Kim S, Kim D, Kim S, So CH, et al. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics. 2019;36:1234–40. .
- 13. Wang J, Kou Y, Zhang Y, Gao N, Tu C. Leveraging knowledge context information to enhance personalized recommendation. Neural Inf Process. 2020;467–78.
- 14.
Sanh V, Debut L, Chaumond J, Wolf T. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. https://doi.org/10.1093/bioinformatics/btz682 2019.
- 15. Khan S, Hakak S, Deepa N, Prabadevi B, Dev K, Trelova S. Detecting COVID-19-related fake news using feature extraction. Front Public Health. 2022;9.
- 16.
Wani A, Joshi I, Khandve S, Wagh V, Joshi R. Evaluating deep learning approaches for Covid19 fake news detection. In: Combating Online Hostile Posts in Regional Languages during Emergency Situation. Springer International Publishing; 2021, pp. 153–63.
- 17.
Kazemi SM, Poole D. SimplE embedding for link prediction in knowledge graphs. arXiv, preprint, arXiv:1802.04868. 2018.
- 18.
Akhtar MM, Sharma B, Karunanayake I, Masood R, Ikram M, Kanhere SS. Machine learning-based automatic annotation and detection of COVID-19 fake news. 2022.
- 19.
Patwa P, Sharma S, Pykl S, Guptha V, Kumari G, Akhtar MS, et al. Fighting an infodemic: COVID-19 fake news dataset. Combating Online Hostile Posts in Regional Languages during Emergency Situation. Springer International Publishing; 2021, pp. 21–29.
- 20. Shu K, Sliva A, Wang S, Tang J, Liu H. Fake news detection on social media: a data mining perspective. ACM SIGKDD Explor Newslett. 2017;19(1):22–36. 10.1145/3137597.3137600
- 21. Khanday A, Khan QR, Rabani S. Identifying propaganda from online social networks during COVID-19 using machine learning techniques. Int J Inf Technol. 2020;1–8.
- 22.
Shu K, Cui L, Wang S, Lee D, Liu H. DEFEND: explainable fake news detection. Association for Computing Machinery; 2019; pp. 395–405.
- 23. Nasir JA, Khan OS, Varlamis I. Fake news detection: A hybrid CNN-RNN based deep learning approach. Int J Inf Manage Data Insights. 2021;1(1):100007.
- 24. Ilie V-I, Truic˘a C-O, Apostol E-S, Paschke A. Context-aware misinformation detection: a benchmark of deep learning architectures using word embeddings. IEEE Access. 2021;9:162122–46.
- 25. Truică C-O, Apostol E-S. MisRoBÆRTa: transformers versus misinformation. Mathematics. 2022.
- 26. Truică C-O, Apostol E-S. It’s all in the embedding! Fake news detection using document embeddings. Mathematics. 2023;11. DOI:
- 27. Apostol E-S, Truică C-O, Paschke A. ContCommRTD: a distributed content-based misinformation-aware community detection system for real-time disaster reporting. IEEE Trans Knowl Data Eng. 2024;36:5811–22.
- 28. Truică C-O, Apostol E-S, Panagiotis K. DANES: deep neural network ensemble architecture for social and textual context-aware fake news detection. Knowl-Based Syst. 2024;294.
- 29.
Truică C-O, Apostol ES, Paschke A. Awakened at CheckThat!-2022: fake news detection using bilstm and sentence transformer. In: Conference and Labs of the Evaluation Forum, 2022. Available from: https://api.semanticscholar.org/CorpusID:251471964.
- 30. Raza S, Ding C. Fake news detection based on news content and social contexts: a transformer-based approach. Int J Data Sci Anal. 2022;13:1–28. 10.1007/s41060-021-00302-z.
- 31.
Zhou Z, Guan H, Bhat MM, Hsu J. Fake news detection via NLP is vulnerable to adversarial attacks. In: 11th International Conference on Agents and Artificial Intelligence, 2019.
- 32.
Potthast M, Kiesel J, Reinartz K, Bevendorff J, Stein B. A stylometric inquiry into hyperpartisan and fake news. arXiv, preprint, arXiv:1702.05638, 2017.
- 33.
Kong SH, Tan LM, Gan KH, Samsudin NH. Fake news detection using deep learning. In: 2020 IEEE 10th Symposium on Computer Applications & Industrial Electronics (ISCAIE), 2020, pp. 102–7.
- 34. Zhao J, Zhao Z, Shi L, Kuang Z, Liu Y. Collaborative mixture-of-experts model for multi-domain fake news detection. Electronics. 2023;12:3440.
- 35.
Magdy A, Wanas N. Web-based statistical fact checking of textual documents. In: Proceedings of the 2nd International Workshop on Search and Mining User-generated contents. Association for Computing Machinery; 2010, pp. 103–10.
- 36. Wu Y, Agarwal PK, Li C, Yang J, Yu C. Toward computational fact-checking. Proc. VLDB Endow. 2014;7:589–600.
- 37. Ciampaglia GL, Shiralkar P, Rocha LM, Bollen J, Menczer F, Flammini A. Computational fact checking from knowledge networks. PLOS ONE. 2015;10:1–13. 10.1371/journal.pone.0128193.
- 38. Seddari N, Derhab A, Belaoued M, Halboob W, Al-Muhtadi J, Bouras A. A hybrid linguistic and knowledge-based analysis approach for fake news detection on social media. IEEE Access. 2022;10:62097–109.
- 39. Thilagam PS, et al. Multi-layer perceptron based fake news classification using knowledge base triples. Appl Intell. 2023;53:6276–87.
- 40.
Zhang H, Fang Q, Qian S, Xu C. Multi-modal knowledge-aware event memory network for social media rumor detection. In: Proceedings of the 27th ACM International Conference on Multimedia, 2019, pp. 1942–51.
- 41.
Wu K, Yuan X, Ning Y. Incorporating relational knowledge in explainable fake news detection. In: Pacific-Asia Conference on Knowledge Discovery and Data Mining, 2021, pp. 403–15.
- 42.
Li J, Ni S, Kao H-Y. Meet the truth: leverage objective facts and subjective views for interpretable rumor detection. arXiv, preprint, arXiv:2107.10747, 2021.
- 43. Apostol ES, Coban Ö, Truică C-O. CONTAIN: a community-based algorithm for network immunization. Eng Sci Technol. 2024;55:101728.
- 44. Truică C-O, Apostol E-S, Nicolescu R-C, Karras P. MCWDST: a minimum-cost weighted directed spanning tree algorithm for real-time fake news mitigation in social media. IEEE Access. 2023;11:125861–73.
- 45.
Petrescu A, Truică C-O, Apostol E-S, Karras P. Sparse shield: social network immunization vs. harmful speech. In: Proceedings of the 30th ACM International Conference on Information & Knowledge Management. Association for Computing Machinery; 2021, pp. 1426–36. DOI: 10.1145/3459637.3482481
- 46. Parmar S, Rahul . Fake news detection via graph-based Markov chains. Int J Inf Technol. 2023;16:1333–1345. DOI: https://doi.org/10.1007/s41870-023-01558-3.
- 47.
Shakeel D, Jain N. Fake news detection and fact verification using knowledge graphs and machine learning. 2021.
- 48. Bordes A, Usunier N, Garcia-Duran A, Weston J, Yakhnenko O. Translating embeddings for modeling multi-relational data. Adv Neural Inf Process Syst. 2013;26.
- 49.
Bishan Y, Yih W, He X, Gao J, Deng L. Embedding entities and relations for learning and inference in knowledge bases. In: 3rd International Conference on Learning Representations, ICLR San Diego, CA, USA May 7–9, 2015 Conference Track Proceedings, 2015.
- 50.
Pan JZ, Pavlova S, Li C, Li N, Li Y, Liu J. Content based fake news detection using knowledge graphs. In: The Semantic Web – ISWC 2018 - 17th International Semantic Web Conference, 2018, Proceedings. Springer; 2018; pp. 669–683. https://doi.org/10.1007/978-3-030-00671-6_39.
- 51.
Hu L, Yang T, Zhang L, Zhong W, Tang D, Shi C, et al. Compare to the knowledge: graph neural fake news detection with external knowledge. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Association for Computational Linguistics; 2021, pp. 754–63.
- 52.
Dun Y, Tu K, Chen C, Hou C, Yuan X. KAN: knowledge-aware attention network for fake news detection. In: Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 35, 2021, pp. 81–9.
- 53.
Vaibhav V, Annasamy RM, Hovy E. Do sentence interactions matter? Leveraging sentence level representations for fake news classification. arXiv, preprint, arXiv:1910.12203, 2019.
- 54.
Nguyen V-H, Sugiyama K, Nakov P, Kan M-Y. Fang: leveraging social context for fake news detection using graph representation. In: Proceedings of the 29th ACM international conference on information & knowledge management, 2020, pp. 1165–74.
- 55.
Zhang J, Dong B, Philip SY. Fakedetector: effective fake news detection with deep diffusive neural network. In: 2020 IEEE 36th international conference on data engineering (ICDE), 2020, pp. 1826–9.
- 56. Bian T, Xiao X, Xu T, Zhao P, Huang W, Rong Y, Huang J. Rumor detection on social media with bi-directional graph convolutional networks. Proc AAAI Conf Artif Intell. 2020;34:549–56.
- 57.
Dou Y, Shu K, Xia C, Yu PS, Sun L. User preference-aware fake news detection. In: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, 2021, pp. 2051–5.
- 58. Sargsyan A, Kodamullil AT, Baksi S, Darms J, Madan S, Gebel S, et al. The COVID-19 ontology. Bioinformatics. 2020;36:5703–5.
- 59.
Wang LL, Lo K, Chandrasekhar Y, Reas R, Yang J, Burdick D, et al. CORD-19: the COVID-19 open research dataset. In: Proceedings of the 1st Workshop on NLP for COVID-19 at ACL 2020. Association for Computational Linguistics; 2020.
- 60.
Angeli G, Premkumar MJJ, Manning CD. Leveraging linguistic structure for open domain information extraction. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Association for Computational Linguistics; 2015, pp. 344–54.
- 61. Karnyoto AS, Sun C, Liu B, Wang X. Transfer learning and GRU-CRF augmentation for COVID-19 fake news detection. Comput Sci Inf Syst. 2022;19:639–58.
- 62.
Sharif O, Hossain E, Hoque M. Combating hostility: Covid-19 fake news and hostile post detection in social media. arXiv, preprint, arXiv:2101.03291. 2021.
- 63. Karnyoto AS, Sun C, Liu B, Wang X. Augmentation and heterogeneous graph neural network for AAAI2021-COVID-19 fake news detection. Int J Mach Learn Cybern. 2022;13:2033–2043. Available at: https://api.semanticscholar.org/CorpusID:245832000.
- 64. Biradar S, Saumya S, Chauhan A. Combating the infodemic: COVID-19 induced fake news recognition in social media networks. Complex Intell Syst. 2022;9.
- 65. Alghamdi J, Lin Y, Luo S. Towards COVID-19 fake news detection using transformer-based models. Knowl-Based Syst. 2023;274:110642.
- 66.
Li X, Xia Y, Long X, Li Z, Li S. Exploring text-transformers in AAAI 2021 shared task: COVID-19 fake news detection in English. In: Chakraborty T, Shu K, Bernard HR, Liu H, Akhtar MS, editors. Combating Online Hostile Posts in Regional Languages during Emergency Situation. CONSTRAINT 2021. Communications in Computer and Information Science, vol 1402. Cham: Springer; 2021. https://doi.org/10.1007/978-3-030-73696-5_11
- 67.
Raha T, Indurthi V, Upadhyaya A, Kataria J, Bommakanti P, Keswani V, et al. Identifying COVID-19 fake news in social media. arXiv, preprint, arXiv:2101.11954. 2021.
- 68.
Patwa P, Bhardwaj M, Guptha V, Kumari G, Pykl S, Das A, et al. Overview of CONSTRAINT 2021 shared tasks: detecting English COVID-19 fake news and Hindi hostile posts. In: Chakraborty T, Shu K, Bernard HR, Liu H, Akhtar MS, editors. Combating Online Hostile Posts in Regional Languages during Emergency Situation. CONSTRAINT 2021. Communications in Computer and Information Science, vol 1402. Cham: Springer; 2021. https://doi.org/10.1007/978-3-030-73696-5_5
- 69.
Glazkova A, Glazkov M, Trifonov T. g2tmn at Constraint@AAAI2021: exploiting CT-BERT and ensembling learning for COVID-19 fake news detection. In: Combating Online Hostile Posts in Regional Languages during Emergency Situation. Springer International Publishing; 2021, pp. 116–27.
- 70. Varshney D, Vishwakarma DK. An automated multi-web platform voting framework to predict misleading information proliferated during COVID-19 outbreak using ensemble method. Data Knowl Eng. 2023;143:102103.
- 71. Dipta Das S, Basak A, Dutta S. A heuristic-driven uncertainty based ensemble framework for fake news detection in tweets and news articles. Neurocomputing. 2021;491:607–20.
- 72.
Truică C-O, Leordeanu C. Classification of an imbalanced data set using decision tree algorithms. University Politehnica of Bucharest Scientific Bulletin Series C – Electrical Engineering and Computer Science, Vol. 79, pp. 69, 2017.