Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Detecting nuance in conspiracy discourse: Advancing methods in infodemiology and communication science with machine learning and qualitative content coding

  • Michael Robert Haupt ,

    Roles Conceptualization, Formal analysis, Methodology, Visualization, Writing – original draft, Writing – review & editing

    mhaupt@ucsd.edu

    Affiliations Department of Cognitive Science, University of California San Diego, La Jolla, California, United States of America, Global Health Policy & Data Institute, San Diego, California, United States of America

  • Michelle Chiu,

    Roles Conceptualization, Writing – review & editing

    Affiliation Department of Psychology, Temple University, Philadelphia, Pennsylvania, United States of America

  • Joseline Chang,

    Roles Conceptualization, Project administration

    Affiliation Rady School of Management, University of California San Diego, La Jolla, California, United States of America

  • Zoe Li,

    Roles Data curation, Formal analysis

    Affiliations Global Health Policy & Data Institute, San Diego, California, United States of America, S-3 Research, San Diego, California, United States of America

  • Raphael Cuomo,

    Roles Supervision, Writing – review & editing

    Affiliation Department of Anesthesiology, University of California, San Diego School of Medicine, San Diego, California, United States of America

  • Tim K. Mackey

    Roles Supervision, Writing – review & editing

    Affiliations S-3 Research, San Diego, California, United States of America, Global Health Program, Department of Anthropology, University of California, San Diego, California, United States of America

Abstract

The spread of misinformation and conspiracies has been an ongoing issue since the early stages of the internet era, resulting in the emergence of the field of infodemiology (i.e., information epidemiology), which investigates the transmission of health-related information. Due to the high volume of online misinformation in recent years, there is a need to continue advancing methodologies in order to effectively identify narratives and themes. While machine learning models can be used to detect misinformation and conspiracies, these models are limited in their generalizability to other datasets and misinformation phenomenon, and are often unable to detect implicit meanings in text that require contextual knowledge. To rapidly detect evolving conspiracist narratives within high volume online discourse while identifying nuanced themes requiring the comprehension of subtext, this study describes a hybrid methodology that combines natural language processing (i.e., topic modeling and sentiment analysis) with qualitative content coding approaches to characterize conspiracy discourse related to 5G wireless technology and COVID-19 on Twitter (currently known as ‘X’). Discourse that focused on correcting 5G conspiracies was also analyzed for comparison. Sentiment analysis shows that conspiracy-related discourse was more likely to use language that was analytic, combative, past-oriented, referenced social status, and expressed negative emotions. Corrections discourse was more likely to use words reflecting cognitive processes, prosocial relations, health-related consequences, and future-oriented language. Inductive coding characterized conspiracist narratives related to global elites, anti-vax sentiment, medical authorities, religious figures, and false correlations between technology advancements and disease outbreaks. Further, the corrections discourse did not address many of the narratives prevalent in conspiracy conversations. This paper aims to further bridge the gap between computational and qualitative methodologies by demonstrating how both approaches can be used in tandem to emphasize the positive aspects of each methodology while minimizing their respective drawbacks.

1. Introduction

1.1 The emergence of infodemiology in the early internet era

Since the 2016 US presidential election, online misinformation spread had become a ubiquitous topic in public discourse with rising concerns related to the proliferation of “fake news” on social media platforms such as Facebook and Twitter before the election [13]. Public concern about misinformation spread has only grown since then, with the World Health Organization (WHO) declaring an ‘infodemic’ in reaction to the proliferation of misinformation related to the COVID-19 pandemic [4]. However, risk related to online misinformation has existed since the early stages of the internet, as shown by the emergence of the field of infodemiology during that era [5, 6]. Infodemiology, also known as “information epidemiology,” is the study of the determinants and distribution of health information and misinformation, and identifies areas where there is a knowledge translation gap between best available evidence and what most people do or believe, as well as markers for “high-quality” information [5, 7]. Infodemiological frameworks make distinctions between supply-based applications, such as analyzing what is being published on web sites and social media, and demand-based methods that examine search and navigation behavior on the internet [7]. Applications of infodemiology include the analysis of queries from internet search engines to predict disease outbreaks [6], monitoring peoples’ health status updates on platforms such as Twitter and Weibo for syndromic surveillance [810], identifying prevalent themes and discourse around health conditions and behaviors (including COVID-19) [7, 11, 12], and studies examining online mobilization of social media users to influence health policy outcomes [13, 14].

While social media platforms are an incredible tool for staying connected and communicating with others, they can also simultaneously be a source of uncertainty and fear, which are often accompanied by increased dissemination of unverified rumors, misinformation, and fringe or conspiracy theories [15, 16]. To minimize the spread of misinformation, it is crucial to rapidly identify and then systematically sift through the high volume of posts and comments that often accompanies related online narratives, such as claims stating false ties between COVID-19 and 5G wireless technology. In response to this growing need for rapid content characterization, particularly in the context of health emergencies, supervised and unsupervised machine learning approaches, such as natural language processing (NLP) and supervised learning trained on fact-checked data, have been used to identify and classify misinformation [1720]. Additionally, large language models (LLM) have shown high accuracy scores for identifying misinformation [21, 22], making popular models like chatGPT another promising tool for future infodemiological research. Despite the fact that various machine learning approaches are effective at classifying information and are crucial for adequately addressing the ongoing proliferation of misinformation that is overwhelming the social media ecosystem, these methods can be limited by their associated lexicon or dictionary and the semantic shifts that naturally occur in language over time and across contexts, including changing evidence about COVID-19 and other public health issues that may arise. Thus, for emerging issues that would not be recorded within training datasets, maintaining the use of human review and annotation remains necessary for properly detecting and contextualizing novel situations and narratives.

To gain a more thorough understanding of the nuanced patterns and dynamics underlying the spread of misinformation, our study introduces a hybrid analytic approach using metadata from 5G-COVID conspiracy discourse on Twitter (currently known as ‘X’ but referred to as Twitter for this paper) aimed at leveraging both the efficiency of NLP techniques and the qualitative schema afforded by human coders. More specifically, this study used topic modeling and sentiment analysis to identify influential posts and characterize the discourse, and then used both inductive and deductive coding to detect context-specific narratives that would not be measured by standard sentiment dictionaries. While NLP and qualitative coding methods have been widely used throughout the literature for identifying and characterizing misinformation (see the following for examples: [13, 2325]), this paper combines these methods into a streamlined approach that can be utilized for rapidly characterizing nuanced themes and emerging narratives within large scale online discourses that requires significantly less burden on human annotation, as illustrated in the flow diagram in S1 Appendix. The hybrid method utilized in this study also has the potential to generate nuanced training data for machine learning classifier models without the need for annotating thousands of posts.

This study aims to further bridge the gap between computational and qualitative methodologies by demonstrating how NLP and manual annotation can be used synergistically to emphasize the positive aspects of each approach while minimizing their respective drawbacks. In order to assess the utility of this methodology, themes generated by this approach will be compared for consistency with 5G-COVID conspiracy themes previously identified on Twitter [2628], Facebook [29], and Instagram [30]. Further, this study will extend the current literature by providing in-depth characterization of discourse focusing on correcting false information, which can be used to inform counterstrategies for online misinformation propagation.

1.2 Using machine learning to detect misinformation

Even before the proliferation of misinformation spurred by the emergence of COVID-19, there have been several research efforts over the past decade demonstrating that machine learning approaches can be effective methods for detecting misinformation. Previous work has developed models with at least 70% accuracy for correctly classifying misinformation across multiple platforms and topics, including Twitter posts related to the zika virus [31, 32], YouTube videos about prostate cancer [33], and false health-related information on medical forum posts [34] and Chinese posts on the platforms Weibo and WeChat [35]. Social media users can be classified as well, as shown by Ghenai and Mejova [36], who were able to identify with a 90% accuracy Twitter users prone to sharing false information promoting ineffective cancer treatments.

Many investigations using misinformation classifiers also examine the role of feature selection (i.e., what variables to include in the model) on prediction accuracy. While classifiers can effectively detect misinformation from only using textual features of the post, which are typically keywords associated with specific conspiracy theories or false narratives [37, 38], the inclusion of variables accounting for properties of the post beyond the content of the specific message, such as emotional sentiment, has been shown to improve model performance across multiple studies [17, 18, 20]. Deriving sentiment features from posts only requires basic arithmetic, where sentiment scores are calculated by first counting the number of words associated with a sentiment category and then dividing that count by the total number of words within the post [39]. Each sentiment category includes a dictionary, which is the list of words selected to represent the category. Sentiment dictionaries can measure a wide array of topics, such as emotional affect states (e.g., anger, anxiety, joy), biological processes (e.g., health, death), cognitive processes (e.g., certainty, causation), time orientations (e.g., focus on past, present, or future), and linguistic properties (e.g., number of 1st-person pronouns used) [40].

Assigning words to sentiment dictionaries can be accomplished using data-driven approaches, as shown by Stanford’s Empath project that derived over 200 classification categories from analyzing more than 1.8 billion words of modern fiction [41]. However, for online misinformation research, psychometrically validated sentiment dictionaries provided by the software Linguistic Inquiry and Word Count (LIWC) have been widely used for characterizing misinformation discourse [42, 43] and calculating sentiment features across multiple studies that evaluate performance of misinformation classifiers [4446]. A likely reason sentiment features are effective predictor variables is due to the fact that texts containing misinformation and conspiracies often have an emotional signature such as higher frequency of anger and anxiety words [15, 16, 43], although recent work indicates that emotional valence can vary depending on the type of misinformation being studied [47].

Despite the demonstrated utility of machine learning approaches for static misinformation detection, sentiment analysis and misinformation classifiers are both limited in their ability to adapt to the ever-evolving nature of human language. That is, in the same way that Gestalt psychology demonstrates that the perception of a color shifts depending on what other colors surround it [48], the shades of meanings from words also shift based on the other words surrounding it within a sentence or paragraph. Since word meanings can change due to differences in context-use and language shifts throughout time, results from sentiment scores and classifier models may not always reflect the actual semantic meaning of texts. These limitations can be clearly depicted in other applications such as automated detection models for abusive language as demonstrated by Yin and Zubiaga [49], who show that these classifier models are limited in their generalizability for other abusive language datasets, and may over-rely on the appearance of keywords such as slurs and profanity. While slurs and profanity can be strong predictors of online abusive language, abuse can also be expressed using implicit meanings and subtext, which results in models that overlook abuse in posts not including slurs and profanity. It is also possible that a post containing profanity keywords is not abuse at all, such as instances of teasing between friends, yet a model would falsely label as abuse [49]. Due to the continually evolving nature of language and context-dependent meanings, detecting misinformation and conspiracies only using machine learning models faces similar semantic challenges.

Interpreting sentiment scores can be further complicated when accounting for the fact that the discussion topic can also influence the average emotional tone of a discourse. For example, discussions about a pandemic may have higher percentages of negative affect words (e.g., “death”, “tragedy”) compared to lighter conversation topics such as gardening. Hence, it is difficult to determine what is an appropriate threshold for meaningful sentiment scores across topics. While this limits what researchers can infer from these analyses, this study addresses this limitation by using a relativistic interpretation of sentiment scores. More specifically, discourse will be marked as high in a sentiment category based on whether it is in the 90th percentile of scores within the corpus. Using the 90th percentile as a threshold marker allows us to account for the specific context of the 5G discourse when making judgements for determining what comprises a high level of sentiment. This approach is also adaptable across a wide array of topics since percentiles indicate which cases are high or low for a given metric based on the sample distribution. To further illustrate this point, a post containing 5% of death-related words may be in the 95th percentile for discourse about gardening, making it a “high” amount, but be within the 50th percentile for pandemic-related discourse, making it a typical percentage within the context of that corpus. Additionally, this study incorporates a qualitative coding approach to account for contextual information that is typically overlooked in sentiment analysis but recognizable by a human coder.

1.3 Inductive and deductive approaches for content coding

In recent years there have been calls for researchers to recognize the impact of contextualization when interpreting data, such as accounting for changes in semantics within a dataset over time [50]. When identifying emerging narratives within online discourse, where meanings of text vary due to novel co-occurrences of words, the boundaries between noise and relevant signal are constantly shifting. Compared to supervised machine learning approaches, which requires an existing training dataset of annotated posts to be effective, human coders are more adaptable to updating their background knowledge of events, making them effective at recognizing text containing previously undocumented narratives. To account for contextualization factors, our study utilized both inductive and deductive analytic techniques to examine tweets related to conspiracy discourse stating a relationship between COVID-19 and 5G technology.

Based in grounded theory [51], inductive coding is an iterative data analytic process centered on constant examination and comparison, which allows for theory development and explicit coding procedures [52, 53]. The primary benefits of an inductive approach is that it allows researchers to code texts using labels that are both aligned with the data and free from the influence of extant concepts, as well as detect tacit elements or connotations of the data that may not be apparent from a superficial reading of denotative content [54]. For the current study, inductive coding was used to identify narratives and rumors prominent within the 5G discourse.

Deductive coding, on the other hand, refers to a top-down coding process intended on testing whether data coincide with existing assumptions, theories, or hypotheses [55]. For the deductive coding scheme in the current study, we coded for whether posts contained misinformation or misinformation corrections based on whether it made statements claiming that 5G wireless technology causes COVID-19, or explicitly refuted the conspiracy. The criteria for classifying 5G-related misinformation and corrections were adapted from previous frameworks identifying COVID-related misinformation [5658]. The researchers also coded for whether tweets expressed a positive, negative, or neutral stance towards 5G conspiracies. Coding for the user’s stance makes it possible to trace and compare general sentiment across topics without having to account for specific themes.

By using both inductive and deductive coding schemes in our content analysis, we aimed to extract and interpret the underlying patterns of COVID-19 misinformation in a more comprehensive manner that are both consistent with the textual data and aligned with existing conceptual frameworks. This approach also allows us to track both general conceptual categories (e.g., misinformation) and more case-specific incidents (e.g., religious pastor claiming 5G causes COVID-19), or in other words, seeing both “the forest and the trees.”

1.4 Combining unsupervised machine learning and content analysis approaches

Previous studies that only use content analysis when investigating social media activity can require the coding of hundreds, if not thousands of posts, which can be an extensive cost in time and resources. Fortunately, unsupervised machine learning and NLP approaches have been used to aid in content analysis on social media, which includes the detection of illicit drug sales or promotions [5962], self-report symptoms on social media [63], and identifying prominent themes in COVID-related misinformation [56]. These approaches group posts together based on textual similarity, which can help filter out irrelevant topic clusters based on keywords [64], or organize the posts together to facilitate the identification of higher-level themes [56]. Since most social media discourse is driven by a small number of highly active users while the majority of users typically engage in passive behaviors such as browsing [6568], characterizing discourse based on the most highly shared posts can be an effective approach for assessing public response to an issue due to most users not often generating their own content. This is further demonstrated in a recent study that conducted a social network analysis of Twitter users, which showed that the propagation of a COVID-related conspiracy theory was initiated by only a handful of prominent accounts during the early stages of the pandemic [69].

Another advantage of applying topic modeling to social media data is that it allows for other prominent sub-themes to be identified from the topic clusters that might have been overlooked by just assessing a more general sampling of the most shared posts, such as the most retweeted tweets. This is demonstrated in previous work by the authors [56] who were able to identify themes related to the misuse of scientific authority within Twitter misinformation discourse after applying biterm topic modeling (BTM) to the data and then coding the top 10 retweeted tweets from each topic cluster. Using this approach, informally referred to as “BTM + 10,” we were able to build on themes identified in their previous characterization of misinformation discourse based on a more general sampling of the top 100 retweeted tweets [58]. While BTM + 10 does not cover the content in every post assigned to a topic, the top 10 most retweeted tweets were able to account for at least 50% of the total tweet volume of each topic cluster within the context of the hydroxychloroquine twitter discourse [56], indicating that influential posts tend to account for a majority volume of tweets within a discourse. It is also worth noting that uncoded tweets are still assigned to a topic cluster based on textual similarity, which increases the likelihood that it would be similar in message content to the most influential posts that were assigned to the same cluster. Increasing the number of coded posts from the 10 most retweeted tweets to 15 or 20 can also account for greater tweet volume without adding substantial burden on the coders. In the current study, we wish to build on the BTM + 10 methodology as carried out in previous work [56] by incorporating sentiment analysis when characterizing 5G-COVID conspiracy discourse on Twitter. Themes identified from previous studies examining 5G conspiracy discourse will be compared to assess the validity of the methodology proposed in the current study. For readers interested in adapting this methodology, see the flow chart diagram in S1 Appendix that further outlines the current approach.

1.5 5G conspiracy theories during the COVID-19 pandemic

Beginning in early April 2020, there had been reports that telecom engineers were facing verbal and physical threats, and that at least 80 mobile towers had been burned down in the United Kingdom (UK), actions fueled by false conspiracy theories blaming the spread of COVID-19 on 5G wireless signals [70]. In order to increase understanding of these destructive acts, recent work has shown associations between belief in 5G-COVID conspiracies with states of anger and greater justification of real-life and hypothetical violence alongside greater intent to engage in similar behaviors in the future [71].

Public figures and celebrities who, whether with deliberate malintent or not, share rumors and falsehoods to their large groups of followers [7275] are also involved in 5G-COVID conspiracy discourse, as shown in a recent study characterizing discourse on Facebook that identifies celebrities and religious leaders as 5G-COVID conspiracy propagators [29]. Bruns et al. (2020) also identify prominent 5G conspiracy theories such as claims stating that 5G reduces the ability of the human body to absorb oxygen, and that 5G is related to a complex agenda involving bioengineered viruses and deadly 5G-activated vaccines led by elite figures such as Bill Gates, George Soros, the World Health Organization (WHO), and secret-society organizations like the Illuminati [29]. From these findings, Bruns et al. (2020) conclude that 5G conspiracy theorists had retrofitted the new information emerging about the virus and its effects on human health into pre-existing worldviews, beliefs, and ideologies to further propagate conspiracist narratives [29]. This is consistent with findings from another 5G conspiracy study conducted on Twitter, which shows that the conspiracies were built on existing ideas set against wireless technologies [28]. On Twitter more specifically, researchers found that videos played a more crucial role in 5G rumor propagation than posts [28], and other work has examined how spatial data has been misconstrued by conspiracists, as shown with the promotion of maps that assert false correlations between the distribution of COVID-19 cases and installations of 5G towers [27].

Another study that used social network analysis found that influential accounts tweeting 5G-COVID conspiracies tended to form a broadcast network structure resembling the structure most typical for accounts from mainstream news outlets and celebrities that are frequently retweeted [26]. A content analysis from this study also shows that over a third of randomly sampled tweets contained views claiming that COVID and 5G were linked, and that there was a lack of an authority figure who was actively combating said misinformation [26]. In order to examine the authors of influential tweets within 5G discourse, the current study will also content code affiliations from the most retweeted user accounts based on publicly available profile data.

2. Methods

2.1 Data collection and analysis overview

A total of 256,562 tweets were collected from the public streaming Twitter API per the terms available at the time of the study using keywords “5G” and covid-related words such as “coronavirus”, “covid-19” between March 25th and April 3rd 2020. We chose this time frame as it represents a period when the 5G conspiracy theory first became prominent, as shown in the spike in volume for “5G” posts in Fig 1. All personal identifiable information from tweets was removed in the reporting of the results to preserve anonymity. We note that due to the change in ownership, API policies, and name of the platform (Twitter has been renamed “X”), the terms and conditions of the streaming API used for data collection for this study are no longer applicable for current studies. IRB approval was not required as all data collected in this study was available in the public domain and results from the study have been deidentified and anonymized. The dataset and R syntax used to generate the results can be found at the following Open Science Framework (OSF) link: https://bit.ly/5G_Conspiracies.

thumbnail
Fig 1. Number of Tweets with keyword “5G” through time highlighting timeframe of current study.

https://doi.org/10.1371/journal.pone.0295414.g001

To analyze the relatively large volume of tweets collected in this study, we used the biterm topic model (BTM), an unsupervised machine learning approach using natural language processing (NLP) further described in the next section, to extract themes from text of tweets as used in prior studies examining COVID-19 topics on social media [10, 56, 63, 64]. The top 10 most retweeted tweets associated with each topic cluster (i.e., the “BTM + 10” approach previously mentioned) were coded using a deductive coding scheme adapted from previous COVID-19 misinformation work [56, 58] to classify posts on whether they contain misinformation, or a misinformation correction (further discussed in Section 2.3). For tweets that may not contain misinformation but still support the notion that 5G signals cause COVID-19 infections, positive and negative stances were coded to assess whether a tweet supported or opposed the conspiracy (Table 1). While the metric of stance is traditionally referred to as “sentiment” in related social media research, this current study will refer to it as “stance” to prevent potential confusion with NLP sentiment analysis metrics further described in the next section. An inductive coding scheme was also used to characterize themes and narratives for both misinformation and correction discourses.

thumbnail
Table 1. Examples of content coded tweets (paraphrased and redacted to retain anonymity).

https://doi.org/10.1371/journal.pone.0295414.t001

2.2 Biterm Topic Model (BTM)

Unsupervised topic modeling strategies, such as BTM, are methods particularly well suited for sorting short text (such as the 280-character limit for tweets) into highly prevalent themes without the need for predetermined coding or a training/labelled dataset to classify specific content. This is particularly useful in characterizing large volumes of unstructured data where predefined themes are unavailable, such as in the case of emerging social movements, novel disease outbreaks, and other emergency events where information changes rapidly [10, 56, 60, 61, 63, 64, 76, 77]. The corpus of tweets containing the 5G keywords was categorized into highly correlated topic clusters using BTM based on splitting all text into a bag of words and then producing a discrete probability distribution for all words for each theme that places a larger weight on words that are most representative of a given theme [78]. While other NLP approaches use unigrams or bigrams for splitting text, BTM uses ‘biterms’, which is a combination of two words from a text (e.g., the text “go to school” has three biterms: ‘go to’, ‘go school’, ‘to school’) and models the generation of biterms in a collection rather than documents [79].

BTM was used for this study because biterms directly model the co-occurrence of words, which increases performance for sparse-text documents such as tweets. Conducting BTM analysis is done initially by setting the BTM topic number (k) and “n” words (for the first round of analysis we set at k = 10, n = 20 to cover several possible misinformation topics that might be present in the corpus). A coherence score is then used to measure how strong the top words from each topic correspond to its respective topic. For this study, the model with k = 20 was chosen because it had the highest coherence score compared to other iterations tested. All data collection and processing were conducted using the programming language Python.

2.3 Deductive and inductive coding schemes

In order to characterize highly prevalent misinformation and conspiratorial narratives in the corpus, the top 10 most retweeted tweets from all 20 BTM topic outputs were extracted and manually coded for relevance first using a deductive coding scheme adapted from existing COVID-19 misinformation themes from the literature [5658], and then coded again using an inductive approach identifying context specific themes related to 5G. While misinformation and conspiracies are distinct concepts, the current study will refer to both as ‘misinformation’ within the analysis for brevity.

In total, 200 unique tweets were reviewed by coders. Tweets were classified as misinformation if they contained declarative statements claiming that 5G causes COVID-19, or statements supporting the conspiracy from sources that convey scientific authority such as from medical experts, scientists, or scientific studies [56]. A tweet was considered a misinformation correction if it explicitly opposes the 5G conspiracy and provided information countering the claims about 5G causing COVID-19. Tweets were also annotated for stance in relation to 5G conspiracy theories, with 1 indicating positive stance, -1 indicating negative stance, and 0 if the tweet only reports information about 5G without stating an opinion, exhibits neutral user sentiment, or is not directly related to the 5G COVID conspiracy discourse.

An inductive coding approach was then used to sub-code for reoccurring themes and narratives associated with each topic cluster that is unique to 5G-related misinformation (see Table 2 below for a description of each identified sub-theme and example tweets). Of the total 200 tweets used for inductive coding, the first and third author divided the sample in half to identify themes. Once theme labels were generated by each coder, both coders met to compare inductive coding labels and combine overlapping themes. All 200 tweets were then reviewed again to classify them based on the finalized coding scheme. Using this approach, tweets can be categorized with the same theme label from inductive coding but classified as misinformation or a correction based on the deductive framework. For example, a tweet labeled as misinformation and the inductive coding theme causative explanation (see Table 2 for description) indicates that the tweet is making a causal claim about a correlation between COVID-19 and 5G. However, a tweet labeled as a misinformation correction but labeled as the same theme indicates that the tweet is making causal claims refuting correlations between COVID-19 and 5G technology. An advantage of this approach is that it showcases overlap in discourse themes between misinformation and correction topic clusters.

2.4 Content coding of Twitter account profiles

Twitter profiles of accounts that produced the top 10 most retweeted tweets for each BTM cluster output were content coded to investigate publicly self-reported occupation among these influential users who were active in the Twitter 5G conspiracy discourse during the study period. Publicly available metadata from user account profiles were retrieved and coded to determine whether descriptions stated that they were a medical doctor or scientist, a religious leader, a government official, or affiliated with the media (e.g., a journalist, TV, or radio personality).

2.5 Calculating LIWC sentiment scores

Sentiment scores were calculated using Linguistic Inquiry and Word Count (LIWC) and reflect the percentage of words within a post that correspond to a given sentiment category. The sentiment scores were calculated to assess language related to: Analytic thinking (metric of logical, formal thinking), Clout (language of leadership, status), Authenticity (perceived honesty, genuineness), and Netspeak (internet slang), cognitive-processes related to information evaluation (i.e., Causation, Discrepancy, Tentative, Certitude), Emotional Affect (i.e., Affect, Positive Tone, Negative Tone, Emotion, Positive emotion, Negative emotion, Anxiousness, Anger, Sadness), social- and health-related topics (i.e., Prosocial behavior, Interpersonal conflict, Moralization, Health, Illness, Death, and Risk), and time-related sentiments (i.e., Time, Past focus, Present focus, Future focus). Since some tweets receive a much greater proportion of engagement within a discourse than others, average sentiment scores for each topic cluster were also weighted based on number of retweets received. This allowed us to characterize emotional sentiment of a topic cluster based on the most influential posts, rather than by posts that received low exposure. Due to the lack of a normal distribution of sentiment scores, we applied a Kruskal-Wallis nonparametric approach and Dunn’s post-hoc test to detect statistically significant differences across BTM topic clusters, with a Bonferroni Correction to account for multiple comparisons.

3 Results

3.1 Content analysis and characterization

From the 256,562 tweets analyzed in this study, Table 3 shows all BTM topics generated and assessed. In this study, the majority of topics have at least 50% of their tweet volume composed of the top 10 tweets with highest volume of retweets, indicating that the top 10 retweets characterize a substantive volume of tweets assigned to each topic. There were 4 topics below 50%, with the lowest percentage being 40.9%. While this still accounts for a substantive percentage of tweet volume, it should be noted that there is more uncertainty associated with the characterizations based on the metrics derived from the top 10 most retweeted tweets for these topic clusters. Topics were further classified as high in misinformation or high in misinformation corrections if the topic had at least 33% of top 10 retweeted tweets associated with the respective categories. Topics were also labeled high in both (i.e., Mixed) if it contained at least 33% of misinformation and corrections, and low for the remaining topics. A threshold of 33% was chosen for this analysis since it indicates a substantive number of tweets for each topic. Out of the 20 topics in this study, 9 were classified as high in misinformation, 5 as high in misinformation corrections, 2 as high in both misinformation and corrections, and 4 as low in both. The topic clusters low in both were not chosen for further analysis.

Table 4 shows each topic cluster with the themes identified from inductive coding, as well as a list of LIWC sentiment categories in which the topic scores in the 90th percentile. Most of the topic clusters score in the 90th percentile for at least one sentiment category, indicating that the topic clusters cover a wide range of emotional sentiments. The results in Table 4 show that themes were spread across topics, with sentiment varying across the discourse as well. While some topic clusters are predominately associated with one theme, as seen in topics 7, 8, 11, and 12, other topic clusters have multiple themes. The misinformation topic with the highest number of themes is topic 10, which includes themes of causative explanations, medical authorities, negative health effects, and video. Within the correction discourse, medical authority was paired with themes related to 5G towers on fire and causative explanations in topics 3 and 18. Among the topic clusters that were predominately misinformation, none of the correction tweets corresponded to any of the inductive coding themes. This suggests that misinformation discourses are closed off and less likely to have involvement from those providing corrections. However, this is not consistent with topic clusters high in corrections, as shown in topic 1 where corrections that include causative explanations are in the same discourse as anti-vax misinformation tweets, and topic 9, which has both misinformation and correction tweets that mention religious figures. These results could indicate misinformation discourse is more likely to interfere with correction discourse. The correction discourse also does not mention themes for Tech Corr (i.e., claims that wireless technology advancements are correlated with disease outbreaks), Elite (i.e., claims that figures like Bill Gates and Elon Musk are promoting COVID-19 with 5G technology), Huawei (i.e., the UK government pulling out a contract from a Chinese communications company due to health concerns over 5G), anti-vax (i.e., claims about future COVID-19 vaccines having adverse health effects) and Geo (i.e., maps showing overlapping distribution between COVID cases and 5G cellular towers), indicating that the corrections are not addressing specific misinformation narratives.

thumbnail
Table 4. Inductive coding themes and LIWC sentiment scores across topics.

https://doi.org/10.1371/journal.pone.0295414.t004

Table 5 groups each topic cluster together based on misinformation classification and shows the percentage of coded tweets corresponding to each inductive coding theme. Among misinformation topics, the themes Huawei (33.5%), video (14.1%), and anti-vax (9.9%) were the three most prominent. The three most prevalent themes for corrections were medical authority (22.1%), causative explanations (10.3%), and celebrities (10.1%). The difference in prevalent themes further illustrates that the two conversations focused on differing conversation topics. Within misinformation discourse, the conspiracy theme Huawei claims that the UK government pulled out of a 5G contract due to concerns about it causing COVID, while the actual circumstance was due to concerns about surveillance from the Chinese government. In this case, a situation with a kernel of truth was misconstrued to perpetuate a false claim. The other themes video and anti-vax indicate conversations occurring off Twitter and plant the seed for future distrust in medical efforts. While the correction discourse made general refutations against the false correlation between 5G and COVID-19, it mostly did not address the specific narratives that were prominent in misinformation discourse, which could mitigate the effectiveness of the corrections. For mixed topic clusters, religious figures (29.8%), medical authority (20.8%), and geographic comparisons (9.9%) were the most common themes. It is possible that themes such as religious figures could be associated with conflicts between people’s faith and public health guidelines, which may contribute to the high volume of both misinformation and corrections. It is also possible that the appearance of medical authorities both supporting and refuting claims that 5G causes COVID-19 could lead to ambiguity and confusion among users.

thumbnail
Table 5. Themes from inductive coding by misinformation classification.

https://doi.org/10.1371/journal.pone.0295414.t005

3.2 LIWC sentiment scores

LIWC sentiment scores were also averaged across BTM topics based on misinformation classifications as shown in Table 6 and visualized as bar graphs in Figs 26. Topics with low amounts of misinformation and corrections were kept in this analysis for comparison. Sentiment categories listed in Table 6 that had statistically significant differences across all misinformation classifications are marked with an asterisk (*). Among Text Analytic sentiments (i.e., Analytic, Clout, Authentic, and Netspeak), misinformation topics on average scored higher than corrections across all categories except Authentic, with these differences being statistically significant (p < .001). These differences are most pronounced when comparing Analytic (78.74% misinformation vs 67.66% correction) and Clout sentiments (46.57% vs 41.45%), indicating that misinformation topics were more likely to use words reflecting logic and social status compared to corrections. The mixed BTM topics that contained high levels of both misinformation and corrections scored second highest in Analytic (77.23%) and Clout (46.35%) sentiments but lowest in Authentic (18.24%) against all other classifications. In contrast, topics that contained low amounts of both misinformation and corrections on average scored the lowest in Analytic (59.06%) and Clout (29.54%), but highest in Authentic (27.06%) and Netspeak (6.44%).

thumbnail
Fig 2. Average percentage of text analytic sentiment across misinformation classifications.

https://doi.org/10.1371/journal.pone.0295414.g002

thumbnail
Fig 3. Average percentage of cognitive sentiment across misinformation classifications.

https://doi.org/10.1371/journal.pone.0295414.g003

thumbnail
Fig 4. Average percentage of affect and emotional sentiment across misinformation classifications.

https://doi.org/10.1371/journal.pone.0295414.g004

thumbnail
Fig 5. Average percentage of social and health-related sentiment across misinformation classifications.

https://doi.org/10.1371/journal.pone.0295414.g005

thumbnail
Fig 6. Average percentage of time-related sentiment across misinformation classifications.

https://doi.org/10.1371/journal.pone.0295414.g006

thumbnail
Table 6. Average LIWC sentiment scores (i.e., percentage of words corresponding to each sentiment category) across misinformation classifications.

https://doi.org/10.1371/journal.pone.0295414.t006

For sentiments reflecting cognitive processes (i.e., Causation, Discrepancies, Tentativeness, and Certitude), corrections on average scored significantly higher than misinformation topics across all sentiment categories, and was highest in Causation words (2.06%) compared to all other topic classifications (p < .001). Topics low in both misinformation and corrections scored highest in Discrepancy (1.98%), Tentative (2.54%), and Certitude (2.02%) sentiments.

Among sentiment for general Affect, misinformation topics on average score significantly higher than corrections (7.03% vs 6.95%, p < .001), and are also higher in general Emotion (5.26% vs 4.92%, p < .001), Negative Emotion (5.15% vs 4.96%, p < .001), and Negative Tone (6.17% vs 5.90%, p < .001). However, misinformation scores lower than corrections in Positive Tone (0.67% vs 0.82%, p < .001), Positive Emotion (.06% vs .14%, p < .001), and Anger (0.08% vs 0.18%, p < .001). There are also statistical differences between misinformation and correction topics for Anxiousness (0.063% vs .058%, p = .023) and Sadness (0.01% vs 0.02%, p < .001), although the magnitude of the differences is fairly small.

For sentiment categories measuring social and health-related topics, correction topics on average scored higher than misinformation topics for Prosocial (0.29% vs 0.18%, p < .001), Moral (0.32% vs 0.24%, p < .001), Illness (4.21% vs 4.18%, p < .001), and Risk (0.23% vs 0.16%, p < .001) sentiments. Misinformation topics scored higher on Conflict (.26% vs .24%, p = .013), Health (4.80% vs 4.62%, p < .001), and Death (0.33% vs 0.12%, p < .001) related words compared to corrections. When compared against the averages of all other misinformation classifications, the mixed topic scored highest in Moral (0.48%) and Risk (0.52%) sentiments.

Lastly, among time-related sentiment categories misinformation topics on average scored higher than all other classifications for general Time (2.66%) and Past-focused (3.11%), but scored lowest for Present-focused (3.59%) and lower than corrections for Future-focused (0.53% vs 0.58%, p < .001) sentiments. Topics low in both misinformation and corrections scored the highest (7.16%) for Present-focused words compared against all other classifications (p < .001).

3.3 Analysis of user account affiliations

Accounts that composed the top ten most retweeted tweets associated with each topic cluster were grouped together by misinformation classification type, as seen in Figs 7 and 8. Accounts were classified as a spreader of misinformation or corrections based on whether they posted tweets from the respective categories as labeled from deductive coding. See S1 Table for account information broken out by each topic cluster. Accounts that were engaged in misinformation discourse were the most likely to be suspended (36.6%) or deleted (18.3%) after two years (the time of analysis) from posting their tweets compared to correction discourse where only 2.8% of accounts were suspended. Since a suspended account refers to when the platform kicks off a user’s account, these results reflect actions taken by Twitter towards addressing misinformation discourse.

thumbnail
Fig 7. Most retweeted accounts: Suspended of deleted by misinformation classification.

https://doi.org/10.1371/journal.pone.0295414.g007

thumbnail
Fig 8. Most retweeted accounts: Affiliations by misinformation classification.

https://doi.org/10.1371/journal.pone.0295414.g008

Of the occupational affiliations for most retweeted users, there were 17 medical affiliates / scientists, 27 employed by the media, 7 government officials, and 2 religious leaders. Among accounts engaged in misinformation discourse, there were some occupational affiliations detected with the media (5.6%), medical affiliates / scientists (4.2%), and religious leaders (1.4%). There was no involvement from government officials detected in the misinformation discourse. However, there were relatively more professional affiliations identified in correction discourse where 33.3% of accounts were affiliated with the media, 22.2% were medical affiliates, and 8.3% were government officials. No religious leaders were associated with correction discourse. These findings show that while there were some media and medical affiliates spreading misinformation, the vast majority focused on correcting false information.

4. Discussion

This study collected tweets related to 5G conspiracy theories and applied both NLP and qualitative content coding to characterize a large-scale online discourse. Sentiment analysis results show that misinformation and conspiracy discourse was more likely to use language that is analytic and references social status, death, conflict, and health. This discourse also scored higher in general emotion, negative emotion, and past-orientation. In contrast, discourse that challenged false information was more likely to use language related to cognitive processes, positive emotions, anger, prosocial tendencies, morality, illness, risk, and be future-oriented. These results reveal that correction topics, when compared to misinformation and conspiracy discourse, are more likely to use explanatory words related to causative arguments (e.g., “because”, “how”) and are more concerned with social relations and health-related consequences. The differences in temporal focus between discourses suggest that misinformation narratives evoke past events more often than corrections, whereas the future-oriented language of correction discourse could reflect concerns regarding health-related consequences to 5G-COVID conspiracy theories. These results also reveal that conspiracy-related discourse is more likely to use combative and negative emotional language, and mention extreme consequences more often than corrections (e.g., using death-related words vs illness-related words). The higher degree of negative and emotional language is consistent with previous studies that used sentiment analysis to examine 5G-COVID conspiracies [43, 80], and provides additional insight into rhetorical strategies used in conspiratorial discourse. Further, these findings are consistent with work examining conspiracy spreaders on Twitter more generally, which found significant differences in negative emotion and death-related words compared to those who engage with scientific content [42].

Results from the inductive coding themes were also detected in related work. More specifically, the themes Group of Elites [29, 30, 81], Anti-vax [30, 81], Celebrity [29], Religious Figures [29], Geography Comparisons [27], Videos [82], Technology correlations with disease outbreaks [28], and Negative health effects [29] were all identified in other studies across Facebook, Twitter, and Instagram. The consistency in findings shows the utility of the approach used in the current study, and reveals a pattern of uniformity in conspiracist narratives across social media platforms.

4.1 Using topic modeling to identify discourse for further investigation

This hybrid approach uses topic modeling to identify influential posts within the corpus across multiple topic areas, making it possible to only need to code a small subset of the total volume of tweets in order to detect prominent themes and narratives within both conspiracy and correction discourses. Additionally, segmenting the discourse into topic clusters and then running sentiment analysis makes it possible to detect smaller subsets of tweets within the larger twitter conversation that are high in emotional affect. This can be useful for quickly identifying emotionally charged conversations that could be prioritized for content assessment from public officials or platforms. For instance, since discourse within misinformation topics 0, 6, and 13 are highest in negative emotion, anger and conflict sentiments, as seen in Table 4, these might be of relevance to those interested in tracking or preventing mobilization responses that lead to consequences in the physical world or lead to “offline harm” (such as deciding to burn down a cell tower). When assessing the inductive coding themes related to these topics, results show that themes concerning viral videos, negative health effects, and the Chinese technology company Huawei are associated with these sentiments. This could lead to follow-up analyses examining the users who lead these conversations, and can help prioritize which narratives within a conspiracy discourse might be most harmful to the public, especially in cases such as Huawei that could be linked to xenophobic attitudes towards Chinese and Asian Americans.

Another topic of interest is cluster 8, which is associated with the anti-vax theme and high in health and illness sentiments. Despite the timeframe of the study being late March to early April, where for those in the US and most English-speaking countries COVID-19 was still an emerging pandemic, it is surprising to encounter anti-vax sentiment towards the COVID-19 vaccine almost nine months before the announcement of the first approved vaccine. Topic 16, which has inductive coding themes related to causative explanations and medical authorities, and high in sentiment scores for Anxiousness and Pro-social language, could also be of interest for those looking to identify persuasive strategies used by online conspiracy spreaders. In this discourse, the combination of credibility signifiers reflected in scientific terminology and pro-social language could make messages more persuasive to those seeking certainty and safety during an unknown health crisis or searching for more credible sources of information to fill an existing information gap.

It is also possible to track reactions to misinformation discourse by examining themes prominent in correction topics, where strategies using causative explanations and calls to medical authorities appear to be a common strategy used in corrections. However, the discourse from the corrections did not directly address many of the narratives prevalent in misinformation conversations (e.g., Huawei, Anti-vax) with the most specific talking-points focusing on criticizing celebrities or religious leaders for spreading conspiracies as also detected by Honcharov et al. in a separate study examining anti-vaccination hashtags of public figures on Twitter [82]. Not addressing specific conspiracist narratives could tamper the effectiveness of correction strategies by not clarifying the information gaps that are capitalized on by misinformation spreaders, especially during times of uncertainty. This is particularly important in light of the findings from the affiliations analysis of the most retweeted users, which showed that media figures, medical scientists, and government officials were more likely to be involved in corrections discourse. Since these accounts are more likely to have higher perceived credibility and a greater reach in audience compared to the average user, it is important that emerging conspiracist narratives are identified early in order to design more effective counter-messaging strategies.

4.2 Drawbacks of solely relying on machine learning approaches

Machine learning approaches for detecting misinformation (or classifying categories within online discourse more generally) have continued to evolve over the years. Recent efforts to improve accuracy of misinformation classifiers have found success using theoretical frameworks based on human information processing to guide feature selection [83], combining features from multiple modalities (i.e., text and visual) [84], and including features related to user response to post and message sources [85]. Multiple algorithms have also been evaluated, ranging from traditional logistic regression models and ensemble approaches such as voting and bagging classifiers [86] to advanced models using deep convolutional neural networks and knowledge graphs that use word embeddings instead of requiring feature selection [87, 88].

In response to the high accuracy of machine learning classifiers at detecting misinformation, there have been calls for action in recent years for researchers to make publicly available textual data with labeled fake news articles to build comprehensive training datasets for misinformation classification [8991]. These findings, in addition to the prominence of advanced LLMs such as chatGPT [92, 93], make it tempting to infer that AI technologies can fully address the issue of large-scale content coding without the use of humans. However, as previously discussed, an important drawback of machine learning approaches is that they are generally limited to detecting fake news within specific contexts. Since misinformation topics and conspiracy theories are always evolving, NLP approaches are unable to generalize to novel situations that are not already documented. For example, classifiers developed in response to the misinformation proliferation surrounding the 2016 US presidential election would not be able to adequately detect COVID-related misinformation since the relevant keywords do not correspond across topics. Maintaining up-to-date classifier models would also be difficult to achieve during the COVID-19 pandemic, where scientific consensus and understanding of the virus have shifted multiple times from its early stages, and continues to evolve even over three years after its initial outbreak. It is also difficult for machine learning approaches to detect instances where outdated information or findings that were considered accurate at the time of publication are then used misleadingly in contemporary scenarios, such as in the case surrounding misinformation promoting the use of hydroxychloroquine for treating COVID-19 [56, 58].

While building effective classifier models for 5G COVID conspiracies is possible during 2023 after being well documented, only human coders had the capability to accurately recognize text that contains these narratives when they first emerged online in 2020. We intended to address these limitations by demonstrating the general utility of a hybrid approach that incorporates human coders to take advantage of the efficiencies machine learning techniques provide researchers when working with big datasets while still accounting for contextual nuance from using qualitative approaches. Even though the definition of misinformation may vary in other discussions (e.g., politics, climate change, other social issues), the general principles of the methodology described could be leveraged to provide more up-to-date and richer contextual insights into how these conspiracy-related discourses evolve over time.

4.3 Future directions in infodemiology

In order to properly make sense of findings generated from large scale communication networks, it is important that we also advance our conceptual understandings of information transmission dynamics. Other fields such as cognitive science and human-computer interactions (HCI) have developed frameworks for scientifically examining information as a phenomenon, which can be incorporated into future infodemiology research. One such theory is information foraging, which is based on ecological models of food scavenging behaviors where online users are considered “foragers” who balance the value gained from finding new information with the time cost needed to obtain it [9498]. The types of analyses conducted using an information foraging framework include the following: information patch models that assess engagement activities in environments where information is encountered in clusters (e.g., webpages), information scent models that assess how information value is evaluated from proximal cues (e.g., titles and images on a website), and information diet models that examine decisions concerning the selection and pursuit of information items [94]. Future infodemiological work can design behavioral experiments using information scent models to test how cues on sites or social media posts influence safety perceptions of potentially dangerous transactions such as illicit drug purchasing, and adapt information diet models to examine how users evaluate the truthfulness of health-claims across online environments.

Another relevant theoretical framework is distributed cognition, which extends human intelligence beyond the boundaries of individual actors to encompass interactions between people and resources within the environment [99101]. According to this framework, social organization determines the way information flows through groups [100], and more recent work emphasizes the ways cognition is distributed over a vast array of social, institutional, political, and technological systems that are shaped by, and shape, the individuals who develop and operate within them [99, 102]. Within the context of social media environments, where topics of discourse are constantly shifting and information is transmitted from a wide array of sources including other users, celebrities, institutions, and political figures, this framework can make sense of how narratives evolve by examining the interplay between the broadcaster of a message and their audiences’ reaction, engagement, and retransmission of that message. The theory of distributed cognition can also guide research questions that distinguish between message transmission due to an individual’s personal beliefs and propagation driven by conformity to other users. Overall, both information foraging and distributed cognition frameworks can be adapted in future work to generate deeper insights concerning health-related information seeking behaviors, enrich interpretations from social network analysis, and guide the design of interventions targeting misinformation propagation.

4.4 Limitations

Content coding took place several months after the initial timeframe of the study. While having the time gap allowed us to assess which accounts were deleted or suspended since the initial 5G discourse, we were unable to code for affiliations of those deleted users using publicly available profile data. As stated in the methods section, the top 10 retweeted tweets do not account for every tweet associated within a topic cluster. While the most retweeted tweets account for a substantive proportion of the topic cluster’s tweet volume, and in most cases a majority, there is still some level of uncertainty when characterizing the discourse even if the uncoded tweets share textual similarity to the coded posts. Additional measures such as the use of sentiment analysis, which in this study was applied to the full topic clusters, can also mitigate these concerns since they account for information provided in the uncoded tweets. Finally, though false information can be categorized as “misinformation,” “disinformation,” “mal-information,” and “conspiracy” based on intent and content, this study did not differentiate between these categories, opting to call all false information, regardless of intent, “misinformation.” This lack of differentiation limits the study’s ability to identify potential differences in rhetoric associated with the nuances of false information dissemination.

5. Conclusion

The advancement of communication technologies and the continued emergence of new social media platforms present difficult challenges for researchers looking to investigate the highly dynamic information ecosystem of the 21st century. Fortunately, these same rapid advancements in technology can also be harnessed by researchers as powerful tools to navigate these complex environments. However, as demonstrated in the current paper, the human perspective is equally crucial in this line of work to compensate for the shortcomings that artificial intelligences have toward understanding human endeavors. Due to the rapid pace of modern discourse, words that can be key identifiers for a dangerous conspiracy in one context can be completely irrelevant in a different grouping of text. Within these online conversations, where the boundary between signal and noise is constantly shifting due to emerging and continually evolving narratives, it is crucial to recruit the signal detection capabilities of both machine learning models and human beings to adequately address current and future misinformation challenges now endemic in our global information society.

Supporting information

S1 Checklist. STROBE statement—Checklist of items that should be included in reports of observational studies.

https://doi.org/10.1371/journal.pone.0295414.s001

(DOCX)

S1 Appendix. Hybrid approach for characterizing social media discourse using machine learning and qualitative methodologies.

https://doi.org/10.1371/journal.pone.0295414.s002

(PDF)

S1 Table. Most retweeted accounts associated with each BTM cluster.

https://doi.org/10.1371/journal.pone.0295414.s003

(PDF)

Acknowledgments

The authors would like to thank our collaborators in the Cognitive Media Lab at UC San Diego (cognitivemedialab.ucsd.edu) for the productive discussions about online conspiracy discourse that helped inform the current manuscript.

References

  1. 1. Allcott H, Gentzkow M. Social Media and Fake News in the 2016 Election. J Econ Perspect. 2017;31: 211–36.
  2. 2. Bovet A, Makse HA. Influence of fake news in Twitter during the 2016 US presidential election. Nat Commun. 2019;10: 7. pmid:30602729
  3. 3. Grinberg N, Joseph K, Friedland L, Swire-Thompson B, Lazer D. Fake news on Twitter during the 2016 U.S. presidential election. Science. 2019;363: 374–378. pmid:30679368
  4. 4. Zarocostas J. How to fight an infodemic. The lancet. 2020;395: 676. pmid:32113495
  5. 5. Eysenbach G. Infodemiology: The epidemiology of (mis) information. Am J Med. 2002;113: 763–765. pmid:12517369
  6. 6. Eysenbach G. Infodemiology: tracking flu-related searches on the web for syndromic surveillance. American Medical Informatics Association; 2006. p. 244. pmid:17238340
  7. 7. Eysenbach G. Infodemiology and Infoveillance: Framework for an Emerging Set of Public Health Informatics Methods to Analyze Search, Communication and Publication Behavior on the Internet. J Med Internet Res. 2009;11: e11. pmid:19329408
  8. 8. Cuomo RE, Purushothaman V, Li J, Cai M, Mackey TK. A longitudinal and geospatial analysis of COVID-19 tweets during the early outbreak period in the United States. BMC Public Health. 2021;21: 793. pmid:33894745
  9. 9. Li J, Xu Q, Cuomo R, Purushothaman V, Mackey T. Data Mining and Content Analysis of the Chinese Social Media Platform Weibo During the Early COVID-19 Outbreak: Retrospective Observational Infoveillance Study. JMIR Public Health Surveill. 2020;6: e18700. pmid:32293582
  10. 10. Mackey TK, Li J, Purushothaman V, Nali M, Shah N, Bardier C, et al. Big Data, Natural Language Processing, and Deep Learning to Detect and Characterize Illicit COVID-19 Product Sales: Infoveillance Study on Twitter and Instagram. JMIR Public Health Surveill. 2020;6: e20794. pmid:32750006
  11. 11. Chiou H, Voegeli C, Wilhelm E, Kolis J, Brookmeyer K, Prybylski D. The Future of Infodemic Surveillance as Public Health Surveillance. Emerg Infect Dis J. 2022;28: 121. pmid:36502389
  12. 12. Mackey T, Baur C, Eysenbach G. Advancing Infodemiology in a Digital Intensive Era. JMIR Infodemiology. 2022;2: e37115. pmid:37113802
  13. 13. Haupt MR, Xu Q, Yang J, Cai M, Mackey TK. Characterizing Vaping Industry Political Influence and Mobilization on Facebook: Social Network Analysis. J Med Internet Res. 2021;23: e28069. pmid:34714245
  14. 14. Xu Q, Yang J, Haupt MR, Cai M, Nali MC, Mackey TK. Digital Surveillance to Identify California Alternative and Emerging Tobacco Industry Policy Influence and Mobilization on Facebook. Int J Environ Res Public Health. 2021;18. pmid:34769666
  15. 15. Freckelton QC I. COVID-19: Fear, quackery, false representations and the law. Int J Law Psychiatry. 2020;72: 101611. pmid:32911444
  16. 16. van Prooijen J-W, Douglas KM. Conspiracy theories as part of history: The role of societal crisis situations. Mem Stud. 2017;10: 323–333. pmid:29081831
  17. 17. Ajao O., Bhowmik D., Zargari S. Sentiment Aware Fake News Detection on Online Social Networks. ICASSP 2019–2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 2019. pp. 2507–2511.
  18. 18. Bhutani B., Rastogi N., Sehgal P., Purwar A. Fake News Detection Using Sentiment Analysis. 2019 Twelfth International Conference on Contemporary Computing (IC3). 2019. pp. 1–5.
  19. 19. Kolluri N, Liu Y, Murthy D. COVID-19 Misinformation Detection: Machine-Learned Solutions to the Infodemic. JMIR Infodemiology. 2022;2: e38756. pmid:37113446
  20. 20. Lee CJ, Chua HN. Using Linguistics and Psycholinguistics Features in Machine Learning for Fake News Classification Through Twitter. Springer; 2022. pp. 717–730.
  21. 21. Caramancion K. M. Harnessing the Power of ChatGPT to Decimate Mis/Disinformation: Using ChatGPT for Fake News Detection. 2023 IEEE World AI IoT Congress (AIIoT). 2023. pp. 0042–0046.
  22. 22. Johnson SB, King AJ, Warner EL, Aneja S, Kann BH, Bylund CL. Using ChatGPT to evaluate cancer myths and misconceptions: artificial intelligence and cancer information. JNCI Cancer Spectr. 2023;7: pkad015. pmid:36929393
  23. 23. Rashkin H, Choi E, Jang JY, Volkova S, Choi Y. Truth of varying shades: Analyzing language in fake news and political fact-checking. 2017. pp. 2931–2937.
  24. 24. Massey PM, Kearney MD, Hauer MK, Selvan P, Koku E, Leader AE. Dimensions of Misinformation About the HPV Vaccine on Instagram: Content and Network Analysis of Social Media Characteristics. J Med Internet Res. 2020;22: e21451. pmid:33270038
  25. 25. Zhou C, Li K, Lu Y. Linguistic characteristics and the dissemination of misinformation in social media: The moderating effect of information richness. Inf Process Manag. 2021;58: 102679.
  26. 26. Ahmed W, Vidal-Alaball J, Downing J, López Seguí F. COVID-19 and the 5G Conspiracy Theory: Social Network Analysis of Twitter Data. J Med Internet Res. 2020;22: e19458. pmid:32352383
  27. 27. Flaherty E, Sturm T, Farries E. The conspiracy of Covid-19 and 5G: Spatial analysis fallacies in the age of data democratization. Soc Sci Med. 2022;293: 114546. pmid:34954674
  28. 28. Langguth J, Filkuková P, Brenner S, Schroeder DT, Pogorelov K. COVID-19 and 5G conspiracy theories: long term observation of a digital wildfire. Int J Data Sci Anal. 2022;15: 329–346. pmid:35669096
  29. 29. Bruns A, Harrington S, Hurcombe E. ‘Corona? 5G? or both?’: the dynamics of COVID-19/5G conspiracy theories on Facebook. Media Int Aust. 2020;177: 12–29.
  30. 30. Quinn EK, Fazel SS, Peters CE. The Instagram infodemic: cobranding of conspiracy theories, coronavirus disease 2019 and authority-questioning beliefs. Cyberpsychology Behav Soc Netw. 2021;24: 573–577. pmid:33395548
  31. 31. Sicilia R., Giudice S. L., Pei Y., Pechenizkiy M., Soda P. Health-related rumour detection on Twitter. 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). 2017. pp. 1599–1606.
  32. 32. Sicilia R, Lo Giudice S, Pei Y, Pechenizkiy M, Soda P. Twitter rumour detection in the health domain. Expert Syst Appl. 2018;110: 33–40.
  33. 33. Hou R, Pérez-Rosas V, Loeb S, Mihalcea R. Towards automatic detection of misinformation in online medical videos. 2019. pp. 235–243.
  34. 34. Kinsora A., Barron K., Mei Q., Vydiswaran V. G. V. Creating a Labeled Dataset for Medical Misinformation in Health Forums. 2017 IEEE International Conference on Healthcare Informatics (ICHI). 2017. pp. 456–461.
  35. 35. Liu Y., Yu K., Wu X., Qing L., Peng Y. Analysis and Detection of Health-Related Misinformation on Chinese Social Media. IEEE Access. 2019;7: 154480–154489.
  36. 36. Ghenai A, Mejova Y. Fake cures: user-centric modeling of health misinformation in social media. Proc ACM Hum-Comput Interact. 2018;2: 1–20.
  37. 37. Ajao O, Bhowmik D, Zargari S. Fake news identification on twitter with hybrid cnn and rnn models. 2018. pp. 226–230.
  38. 38. Granik M., Mesyura V. Fake news detection using naive Bayes classifier. 2017 IEEE First Ukraine Conference on Electrical and Computer Engineering (UKRCON). 2017. pp. 900–903.
  39. 39. Tausczik YR, Pennebaker JW. The Psychological Meaning of Words: LIWC and Computerized Text Analysis Methods. J Lang Soc Psychol. 2010;29: 24–54.
  40. 40. Boyd RL, Ashokkumar A, Seraj S, Pennebaker JW. The development and psychometric properties of LIWC-22. Austin TX Univ Tex Austin. 2022; 1–47.
  41. 41. Fast E, Chen B, Bernstein MS. Empath: Understanding topic signals in large-scale text. 2016. pp. 4647–4657.
  42. 42. Fong A, Roozenbeek J, Goldwert D, Rathje S, van der Linden S. The language of conspiracy: A psychological analysis of speech used by conspiracy theorists and their followers on Twitter. Group Process Intergroup Relat. 2021;24: 606–623.
  43. 43. Rains SA, Leroy G, Warner EL, Harber P. Psycholinguistic Markers of COVID-19 Conspiracy Tweets and Predictors of Tweet Dissemination. Health Commun. 2021;38: 21–30. pmid:34015987
  44. 44. Castelo S, Almeida T, Elghafari A, Santos A, Pham K, Nakamura E, et al. A topic-agnostic approach for identifying fake news pages. 2019. pp. 975–980.
  45. 45. Che X, Metaxa-Kakavouli D, Hancock JT. Fake News in the News: An Analysis of Partisan Coverage of the Fake News Phenomenon. 2018. pp. 289–292.
  46. 46. Giachanou A, Ghanem B, Ríssola EA, Rosso P, Crestani F, Oberski D. The impact of psycholinguistic patterns in discriminating between fake news spreaders and fact checkers. Data Knowl Eng. 2022;138: 101960.
  47. 47. Charquero-Ballester M, Walter JG, Nissen IA, Bechmann A. Different types of COVID-19 misinformation have different emotional valence on Twitter. Big Data Soc. 2021;8: 20539517211041280.
  48. 48. Beau Lotto R, Purves D. The empirical basis of color perception. Conscious Cogn. 2002;11: 609–629. pmid:12470626
  49. 49. Yin W, Zubiaga A. Hidden behind the obvious: Misleading keywords and implicitly abusive language on social media. Online Soc Netw Media. 2022;30: 100210.
  50. 50. Poirier L. Reading datasets: Strategies for interpreting the politics of data signification. Big Data Soc. 2021;8: 20539517211029320.
  51. 51. Glaser B, Strauss A. Discovery of grounded theory: Strategies for qualitative research. Routledge; 2017.
  52. 52. Corbin JM, Strauss A. Grounded theory research: Procedures, canons, and evaluative criteria. Qual Sociol. 1990;13: 3–21.
  53. 53. Boyatzis RE. Transforming qualitative information: Thematic analysis and code development. sage; 1998.
  54. 54. Suddaby R. From the Editors: What Grounded Theory is Not. Acad Manage J. 2006;49: 633–642.
  55. 55. Fereday J, Muir-Cochrane E. Demonstrating Rigor Using Thematic Analysis: A Hybrid Approach of Inductive and Deductive Coding and Theme Development. Int J Qual Methods. 2006;5: 80–92.
  56. 56. Haupt MR, Li J, Mackey TK. Identifying and characterizing scientific authority-related misinformation discourse about hydroxychloroquine on twitter using unsupervised machine learning. Big Data Soc. 2021;8: 20539517211013844.
  57. 57. Islam MS, Sarkar T, Khan SH, Mostofa Kamal A-H, Hasan SMM, Kabir A, et al. COVID-19–Related Infodemic and Its Impact on Public Health: A Global Social Media Analysis. Am J Trop Med Hyg. 2020;103: 1621–1629. pmid:32783794
  58. 58. Mackey TK, Purushothaman V, Haupt M, Nali MC, Li J. Application of unsupervised machine learning to identify and characterise hydroxychloroquine misinformation on Twitter. Lancet Digit Health. 2021;3: e72–e75. pmid:33509386
  59. 59. Haupt MR, Cuomo R, Li J, Nali M, Mackey TK. The influence of social media affordances on drug dealer posting behavior across multiple social networking sites (SNS). Comput Hum Behav Rep. 2022;8: 100235.
  60. 60. Mackey TK, Kalyanam J, Katsuki T, Lanckriet G. Twitter-Based Detection of Illegal Online Sale of Prescription Opioid. Am J Public Health. 2017;107: 1910–1915. pmid:29048960
  61. 61. Mackey T, Kalyanam J, Klugman J, Kuzmenko E, Gupta R. Solution to Detect, Classify, and Report Illicit Online Marketing and Sales of Controlled Substances via Twitter: Using Machine Learning and Web Forensics to Combat Digital Opioid Access. J Med Internet Res. 2018;20: e10029. pmid:29613851
  62. 62. Shah Neal, Nali Matthew, Bardier Cortni, Li Jiawei, Maroulis James, Cuomo Raphael, et al. Applying topic modelling and qualitative content analysis to identify and characterise ENDS product promotion and sales on Instagram. Tob Control. 2021;32: e153. pmid:34857646
  63. 63. Mackey T, Purushothaman V, Li J, Shah N, Nali M, Bardier C, et al. Machine Learning to Detect Self-Reporting of Symptoms, Testing Access, and Recovery Associated With COVID-19 on Twitter: Retrospective Big Data Infoveillance Study. JMIR Public Health Surveill. 2020;6: e19509. pmid:32490846
  64. 64. Haupt MR, Jinich-Diamant A, Li J, Nali M, Mackey TK. Characterizing twitter user topics and communication network dynamics of the “Liberate” movement during COVID-19 using unsupervised machine learning and social network analysis. Online Soc Netw Media. 2021;21: 100114.
  65. 65. Benevenuto F, Rodrigues T, Cha M, Almeida V. Characterizing user behavior in online social networks. 2009. pp. 49–62.
  66. 66. Lerman K, Ghosh R. Information Contagion: An Empirical Study of the Spread of News on Digg and Twitter Social Networks. Proceedings of the International AAAI Conference on Web and Social Media. 2010 May 16;4(1):90–7.
  67. 67. Papakyriakopoulos O, Serrano JCM, Hegelich S. Political communication on social media: A tale of hyperactive users and bias in recommender systems. Online Soc Netw Media. 2020;15: 100058.
  68. 68. Van Mieghem P, Blenn N, Doerr C. Lognormal distribution in the digg online social network. Eur Phys J B. 2011;83: 251.
  69. 69. Gruzd A, Mai P. Going viral: How a single tweet spawned a COVID-19 conspiracy theory on Twitter. Big Data Soc. 2020;7: 2053951720938405.
  70. 70. Hamilton IA. 77 cell phone towers have been set on fire so far due to a weird coronavirus 5G conspiracy theory. In: Business Insider [Internet]. [cited 30 Mar 2023]. Available: https://www.businessinsider.com/77-phone-masts-fire-coronavirus-5g-conspiracy-theory-2020-5
  71. 71. Jolley D, Paterson JL. Pylons ablaze: Examining the role of 5G COVID-19 conspiracy beliefs and support for violence. Br J Soc Psychol. 2020;59: 628–640. pmid:32564418
  72. 72. Arora VM, Madison S, Simpson L. Addressing Medical Misinformation in the Patient-Clinician Relationship. JAMA. 2020;324: 2367–2368. pmid:33320231
  73. 73. Bruns A, Harrington S, Hurcombe E. Coronavirus Conspiracy Theories: Tracing Misinformation Trajectories from the Fringes to the Mainstream. In: Lewis M, Govender E, Holland K, editors. Communicating COVID-19: Interdisciplinary Perspectives. Cham: Springer International Publishing; 2021. pp. 229–249. https://doi.org/10.1007/978-3-030-79735-5_12
  74. 74. Calac AJ, Haupt MR, Li Z, Mackey T. Spread of COVID-19 Vaccine Misinformation in the Ninth Inning: Retrospective Observational Infodemic Study. JMIR Infodemiology. 2022;2: e33587. pmid:35320982
  75. 75. Shin I, Wang L, Lu Y-T. Twitter and endorsed (fake) news: The influence of endorsement by strong ties, celebrities, and a user majority on credibility of fake news during the COVID-19 pandemic. Int J Commun. 2022;16: 23.
  76. 76. Blekanov I, Tarasov N, Maksimov A. Topic modeling of conflict ad hoc discussions in social networks. 2018. pp. 122–126.
  77. 77. Schück S, Foulquié P, Mebarki A, Faviez C, Khadhar M, Texier N, et al. Concerns Discussed on Chinese and French Social Media During the COVID-19 Lockdown: Comparative Infodemiology Study Based on Topic Modeling. JMIR Form Res. 2021;5: e23593. pmid:33750736
  78. 78. Kalyanam J, Katsuki T, R.G. Lanckriet G, Mackey TK. Exploring trends of nonmedical use of prescription drugs and polydrug abuse in the Twittersphere using unsupervised machine learning. Addict Behav. 2017;65: 289–295. pmid:27568339
  79. 79. Yan X, Guo J, Lan Y, Cheng X. A biterm topic model for short texts. A biterm topic model for short texts. 2013. pp. 1445–1456.
  80. 80. Gerts D, Shelley CD, Parikh N, Pitts T, Watson Ross C, Fairchild G, et al. “Thought I’d Share First” and Other Conspiracy Theory Tweets from the COVID-19 Infodemic: Exploratory Study. JMIR Public Health Surveill. 2021;7: e26527. pmid:33764882
  81. 81. Erokhin D, Yosipof A, Komendantova N. COVID-19 Conspiracy Theories Discussion on Twitter. Soc Media Soc. 2022;8: 20563051221126052. pmid:36245701
  82. 82. Honcharov V, Li J, Sierra M, Rivadeneira NA, Olazo K, Nguyen TT, et al. Public Figure Vaccination Rhetoric and Vaccine Hesitancy: Retrospective Twitter Analysis. JMIR Infodemiology. 2023;3: e40575. pmid:37113377
  83. 83. Zhao Y, Da J, Yan J. Detecting health misinformation in online health communities: Incorporating behavioral features into machine learning based approaches. Inf Process Manag. 2021;58: 102390.
  84. 84. Singh VK, Ghosh I, Sonagara D. Detecting fake news stories via multimodal analysis. J Assoc Inf Sci Technol. 2021;72: 3–17.
  85. 85. Ruchansky N, Seo S, Liu Y. Csi: A hybrid deep model for fake news detection. 2017. pp. 797–806.
  86. 86. Ahmad I, Yousaf M, Yousaf S, Ahmad MO. Fake News Detection Using Machine Learning Ensemble Methods. Uddin MI, editor. Complexity. 2020;2020: 8885861.
  87. 87. Kaliyar RK, Goswami A, Narang P, Sinha S. FNDNet–A deep convolutional neural network for fake news detection. Cogn Syst Res. 2020;61: 32–44.
  88. 88. Pan JZ, Pavlova S, Li C, Li N, Li Y, Liu J. Content based fake news detection using knowledge graphs. Springer; 2018. pp. 669–683.
  89. 89. Nielsen DS, McConville R. Mumin: A large-scale multilingual multimodal fact-checked misinformation social network dataset. 2022. pp. 3141–3153.
  90. 90. Torabi Asr F, Taboada M. Big Data and quality data for fake news and misinformation detection. Big Data Soc. 2019;6: 2053951719843310.
  91. 91. Zhang X, Ghorbani AA. An overview of online fake news: Characterization, detection, and discussion. Inf Process Manag. 2020;57: 102025.
  92. 92. De Angelis L, Baglivo F, Arzilli G, Privitera GP, Ferragina P, Tozzi AE, et al. ChatGPT and the rise of large language models: the new AI-driven infodemic threat in public health. Front Public Health. 2023;11. Available: https://www.frontiersin.org/articles/10.3389/fpubh.2023.1166120 pmid:37181697
  93. 93. King MR, chatGPT. A Conversation on Artificial Intelligence, Chatbots, and Plagiarism in Higher Education. Cell Mol Bioeng. 2023;16: 1–2. pmid:36660590
  94. 94. Pirolli P, Card S. Information foraging. Psychol Rev. 1999;106: 643.
  95. 95. Pirolli P, Card S. Information foraging in information access environments. 1995. pp. 51–58.
  96. 96. Pirolli P. Rational Analyses of Information Foraging on the Web. Cogn Sci. 2005;29: 343–373. pmid:21702778
  97. 97. Pirolli P. An elementary social information foraging model. 2009. pp. 605–614.
  98. 98. Pirolli P, Fu W-T. SNIF-ACT: A Model of Information Foraging on the World Wide Web. In: Brusilovsky P, Corbett A, de Rosis F, editors. User Modeling 2003. Berlin, Heidelberg: Springer Berlin Heidelberg; 2003. pp. 45–54.
  99. 99. Cash M. Cognition without borders: “Third wave” socially distributed cognition and relational autonomy. Socially Ext Cogn. 2013;25–26: 61–71.
  100. 100. Hollan J, Hutchins E, Kirsh D. Distributed cognition: toward a new foundation for human-computer interaction research. ACM Trans Comput-Hum Interact TOCHI. 2000;7: 174–196.
  101. 101. Rogers Y. A brief introduction to Distributed Cognition. In 1997 [cited 2023 Mar 31]. Available from: https://www.semanticscholar.org/paper/A-brief-introduction-to-Distributed-Cognition-Rogers/5633038212a61652781835c02d681922c1620b48.
  102. 102. Gallagher S. The socially extended mind. Socially Ext Cogn. 2013;25–26: 4–12.