Figures
Abstract
The sudden termination of X’s (formerly Twitter) misinformation policy on November 23, 2022, provides an opportunity to assess the effects of lifting content moderation restrictions on vaccine-related discourse. This study examines changes in the prevalence, thematic composition, and engagement of anti-vaccine discourse following X’s policy shift, analyzing tweets from a seven-day period before and after the policy termination (November 16–30, 2022), excluding the announcement date itself from regression analyses. Using GPT-4o for stance classification, thematic categorization, and stance consistency assessment, with validation through external benchmarks and cross-annotator agreement, we find that anti-vaccine tweets increased significantly post-policy (OR = 1.60, 95% CI: 1.50–1.72), particularly via retweets, suggesting content amplification. Sensitivity analyses excluding highly retweeted content revealed that the policy change was also associated with increased creation of new anti-vaccine content. Thematically, health concerns over vaccination became more prominent, while conspiracy-related and anti-mandate narratives declined in relative prevalence. Stance consistency in quote tweets increased, indicating reinforced ideological alignment in anti-vaccine discourse. These results suggest that content moderation policies may constrain both the volume and amplification of anti-vaccine content, with policy removal associated with rapid shifts in discourse patterns.
Citation: Li Y, Chen T, Zhao Y, Ke W, Pang P, McKay D, et al. (2026) Auditing the impact of social media’s policy shift on anti-vaccine discourse: A large language model-driven empirical study. PLoS One 21(4): e0346568. https://doi.org/10.1371/journal.pone.0346568
Editor: Pierluigi Vellucci, Roma Tre University: Universita degli Studi Roma Tre, ITALY
Received: April 2, 2025; Accepted: March 19, 2026; Published: April 29, 2026
Copyright: © 2026 Li et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: https://doi.org/10.5281/zenodo.19019934.
Funding: The author(s) received no specific funding for this work.
Competing interests: The authors have declared that no competing interests exist.
Introduction
The spread of anti-vaccine discourse on social media plays a crucial role in shaping public perceptions of vaccines and influencing vaccine hesitancy [1,2]. Previous studies have shown that exposure to vaccine misinformation can significantly reduce willingness to get vaccinated [3]. A large share of the misinformation circulating during the pandemic centered on vaccines, spreading doubts about their safety and effectiveness. These narratives contributed directly to growing vaccine hesitancy and public distrust in health guidance. [4,5]. Given the widespread impact of online discussions on public health behaviors [3,6,7], understanding the dynamics of anti-vaccine discourse in response to platform policy changes is of critical importance [8].
Social media platforms have long grappled with the challenge of balancing content moderation and freedom of speech when addressing health misinformation [9]. In response to the pandemic, X implemented a misinformation enforcement policy in 2020 [10,11], introducing content labeling, warning systems, and deplatforming measures to combat misleading claims about COVID-19 vaccines. Similar efforts were undertaken by Facebook [12], Instagram [13], and YouTube [14], which enacted policies to remove or restrict misleading content that contradicted public health guidance [15]. These measures were designed to curb the spread of misinformation and promote credible health information.
However, in November 2022, X, under Elon Musk’s leadership, abruptly terminated its COVID-19 misinformation enforcement policy [16]. Musk, describing himself as a “free speech absolutist,” argued that social media platforms should not remove content that is offensive but still legal, stating that if something is in a “gray area,” it should be allowed to remain [17,18]. While extensive research has examined the implementation of content moderation policies and their effectiveness in reducing misinformation spread, far less is known about the consequences of removing such policies once established. The effects of policy removal may not simply mirror the effects of policy implementation in reverse; understanding these dynamics is crucial for evidence-based platform governance. This policy reversal provides an opportunity to examine how the announcement of policy termination was associated with changes in the prevalence and nature of anti-vaccine discourse on the platform.
In this study, we investigate the short-term impact of X’s policy termination on anti-vaccine discourse. Specifically, we seek to answer the following research questions:
- RQ1: How does the termination of X’s COVID-19 vaccine misinformation policy impact the volume of anti-vaccine tweets?
- RQ2: Does the policy change shift the thematic composition of anti-vaccine discourse, and if so, which narratives became more or less prominent?
- RQ3: How does the policy termination affect the stance consistency of quote tweets within vaccine-related discussions?
To address these questions, we collected a dataset of COVID-19 vaccine-related tweets from November 16–30, 2022. For regression analyses, we compared seven days before (November 16–22) and seven days after (November 24–30) the policy announcement, excluding the announcement date itself (November 23) to isolate pre- and post-announcement periods. Using GPT-4o as an annotation tool, we categorized tweets based on their vaccine stance (pro-vaccine, anti-vaccine, neutral) and assigned the identified anti-vaccine tweets into eight thematic categories [8]. We validated the LLM classifier through external benchmark testing and cross-annotator agreement assessment. We then applied logistic regression models to analyze the changes in anti-vaccine prevalence and topic distribution following the policy termination, with sensitivity analyses to assess the robustness of findings when excluding highly retweeted content.
This study provides empirical insights into the role of content moderation in shaping online vaccine discourse. By leveraging a real-world policy shift as a pre-post observational design, we document changes in anti-vaccine discourse following X’s policy termination. In our analysis, we examine three key dimensions: (1) amplification, the degree to which anti-vaccine content spreads through retweets; (2) thematic composition, the specific narratives expressed in anti-vaccine discourse; and (3) stance consistency, whether users quoting vaccine-related tweets express agreement or disagreement with the original content. Our findings have implications for platform regulation, misinformation governance, and the broader intersection of free speech and public health policies.
Related work
Anti-vaccine discourse on social media
Social media has become a major battleground for vaccine-related discourse, with anti-vaccine rhetoric proliferating across platforms. While pro-vaccine content generally dominates public health messaging, anti-vaccine narratives have been found to engage audiences more effectively due to their emotive and polarizing nature [19]. Research has identified several recurring themes within anti-vaccine discourse, including concerns about vaccine safety, distrust of pharmaceutical companies and government institutions, and the belief in personal health autonomy [20,21]. In addition, conspiracy theories, such as claims that vaccines are used for population control or contain microchips, have played a substantial role in fueling skepticism, particularly during the COVID-19 pandemic [22,23]. Despite lacking scientific support, these narratives have gained traction through social media mechanisms that facilitate their amplification within vaccine-hesitant communities.
Recent studies indicate that vaccine misinformation does not need to be factually incorrect to influence hesitancy. Allen et al. [24] found that the most influential content was not outright falsehoods but rather vaccine-skeptical narratives that raised concerns without making verifiably false claims. Broniatowski et al. [25] similarly found that COVID-19 misinformation was less likely to spread than accurate content, suggesting that amplification mechanisms rather than inherent virality drive misinformation visibility.
Social media platforms facilitate the rapid spread of anti-vaccine content through algorithmic amplification and community clustering [26,27]. Studies indicate that engagement-driven ranking systems amplify misleading yet technically accurate narratives [28,29], while automated accounts and coordinated networks systematically amplify skeptical narratives [28]. Hwang et al. [30] identified distinct thematic clusters in vaccine discourse on Twitter, providing a framework for analyzing anti-vaccine narratives.
While much research has analyzed the spread and impact of anti-vaccine discourse, fewer studies have examined how changes in content moderation policies influence the evolution of these narratives. The sudden removal of X’s COVID-19 misinformation policy presents an opportunity to assess how the withdrawal of content moderation affects the prevalence, themes, and engagement patterns of anti-vaccine rhetoric.
Content moderation and policy changes in shaping online discourse
Content moderation policies play a crucial role in controlling the spread of health misinformation. Previous research has demonstrated that interventions such as labeling, fact-checking, and deplatforming can effectively reduce engagement with misinformation [9,31]. For instance, content labeling reduces the perceived credibility of misinformation, while fact-checking interventions mitigate belief persistence [32,33]. However, the effectiveness of content moderation remains debated. While some scholars argue that removing misleading content reduces public exposure to harmful narratives [15], others suggest that excessive moderation may push misinformation to alternative platforms with weaker oversight [34,35].
X’s 2020 COVID-19 misinformation policy, along with similar initiatives by Facebook and YouTube, demonstrated a platform-wide commitment to mitigating public health misinformation [11,12]. However, concerns over freedom of expression have led to ongoing debates about the balance between regulation and free speech [9]. Past research has shown that deplatforming efforts can significantly disrupt the spread of misinformation [36], as seen in X’s removal of QAnon-related content, which reduced the influence of key conspiracy theorists [37]. Conversely, platforms with relaxed moderation policies, such as Parler and Gab, have experienced surges in extremist and misleading content [38].
Algorithmic amplification remains a challenge for misinformation control, as platforms struggle to prevent users from being funneled into misinformation pathways [39]. Similarly, research on COVID-19 misinformation has examined how moderation efforts shape discourse trends [8]. Lanier et al. [40] analyzed COVID-19 disinformation hashtags such as #scamdemic and #plandemic on Parler, demonstrating how vaccine-skeptical narratives proliferate on platforms with minimal moderation. However, relatively few studies have analyzed the effects of removing misinformation policies after they have been established. Mitts et al. [41] found that the removal of anti-vaccine groups on Facebook led to increased anti-vaccine rhetoric on X, highlighting the potential for cross-platform spillover effects. Additionally, Broniatowski et al. [42] showed that Facebook’s misinformation removal efforts did not always decrease overall engagement with anti-vaccine content, as users adapted to platform restrictions. Hobbs et al. [43] examined the sharing of low-credibility URLs on Twitter during reporting of blood clots following the AstraZeneca vaccine, showing how health scares can trigger spikes in misinformation sharing.
Existing research has primarily focused on the implementation of moderation policies rather than their removal. While numerous studies have examined how introducing content moderation reduces misinformation spread [9,15,31], empirical evidence on policy removal remains scarce. The few exceptions [41,42] focus on specific interventions rather than wholesale policy termination. The sudden termination of X’s COVID-19 misinformation enforcement policy provides a pre-post observational design to assess how policy withdrawal was associated with changes in misinformation dynamics. This study contributes to this gap by systematically analyzing the short-term consequences of reversing content moderation efforts, with implications for both platform governance and public health outcomes.
Materials and methods
Data collection
To analyze the changes following X’s policy shift on anti-vaccine discourse, we utilized an existing dataset of COVID-19-related tweets provided by [44]. Although the dataset description was published in 2021, the data collection was continuously maintained and updated through November 2022, covering the period of our study. This dataset, originally collected through Twitter’s Academic API v2, contains tweet IDs that were made publicly available for research purposes. Following X’s data access policies, we rehydrated these tweet IDs to obtain the full text of the corresponding tweets.
To specifically focus on vaccine-related discussions, we applied an additional keyword filtering step inspired by [45]. After retrieving the full dataset, we extracted tweets containing vaccine-related keywords such as “mRNA vaccine,” “booster shot,” and vaccine manufacturer names (for example, “Pfizer” and “Moderna”). We show a few examples in Table 1. This two-step filtering approach ensured that our dataset retained a broad range of COVID-19 discourse while allowing us to analyze vaccine-related narratives in response to X’s policy change.
The collected tweets were categorized into three distinct types, as shown in Table 2. Original tweets (TW) are those composed and posted by users, quote tweets (QT) include a reshared tweet with an added comment, and retweeted tweets (RT) are simply reshared without any modifications.
To improve data quality, we applied preprocessing steps following [44], including deduplication, language filtering (retaining only English tweets), and removal of URLs and redundant whitespace. After preprocessing, we performed stratified sampling based on tweet type and tweet date to ensure a balanced representation across different categories.
Unlike prior studies that often exclude retweets to focus on original content, we retain retweeted tweets to examine the prevalence of anti-vaccine discourse in response to X’s policy shift. This design choice reflects our focus on the information environment to which users are exposed: retweets contribute to the overall concentration of anti-vaccine content in users’ feeds, which is relevant to public health outcomes regardless of content originality. The inclusion of retweets allows us to analyze how anti-vaccine narratives propagate within the platform and assess the patterns of content sharing following the removal of content moderation policies. To address potential concerns about highly viral content driving results, we also conduct sensitivity analyses excluding tweets with ≥ 50 retweets.
Data annotation method
We employed the GPT-4o API for automated tweet annotation, leveraging the Chain-of-Thought prompting technique to enhance the reasoning capabilities of the LLM. Each annotated tweet includes not only a categorical label but also the LLM’s reasoning process and confidence score to ensure classification transparency and robustness.
All three classification tasks, vaccine stance classification, anti-vaccine theme categorization, and stance consistency analysis, were conducted using GPT-4o. The LLM was prompted with task-specific instructions and example-based guidance to improve classification performance across different categories.
Vaccine stance classification
Tweets were categorized into four stance labels: pro-vaccine, anti-vaccine, neutral, and no stance mentioned, as detailed in Table 3. The classification was performed based on the textual content of each tweet, considering explicit statements, implied meanings, and context.
For retweets (RT), we assume that the poster’s vaccine stance aligns with the original tweet, based on prior research [46], which indicates that retweeting behavior generally signifies endorsement rather than opposition. For quote tweets (QT), stance determination requires contextual reasoning by analyzing both the quoted tweet and the poster’s comment. Given the complexity of vaccine discourse, we utilize GPT-4o’s advanced language comprehension capabilities to ensure accurate classification.
Anti-vaccine theme categorization
To examine whether policy termination influenced the structure of anti-vaccine discourse, we classified anti-vaccine tweets into eight thematic categories using GPT-4o, following the topic taxonomy established by Poddar et al. [8]. These eight themes were originally derived through an iterative process combining Latent Dirichlet Allocation for initial topic discovery, human expert consolidation of candidate topics with seed words, and Labelled-LDA [47] for refined topic assignment. In our study, we directly applied these predefined themes as classification labels and employed GPT-4o to assign each anti-vaccine tweet to the most relevant category. Table 4 presents the eight anti-vaccine theme categories with definitions and examples.
Stance consistency analysis
For quote tweets (QT), stance consistency was assessed by comparing the poster’s comment with the quoted tweet. Each QT was classified into one of three categories: consistent, inconsistent, or unclear, as shown in Table 5. GPT-4o performed this classification based on context-aware reasoning, analyzing both the quoted tweet and the poster’s added comment. Since social media language often includes implicit messaging, sarcasm, and figurative speech, LLM-based analysis may encounter challenges in determining stance consistency. In cases where clear intent could not be inferred from the text, the tweet was labeled as unclear.
Annotation validation
To assess the reliability of GPT-4o classification, we conducted two validation procedures.
External benchmark validation. We tested our stance classification prompt on the CAVES dataset [48], a COVID-19 vaccine stance benchmark developed by the same research team that established the thematic taxonomy we adopted [8]. CAVES contains 1,977 expert-labeled tweets with stance annotations. Our classifier achieved F1 = 0.883 for anti-vaccine stance identification (Precision = 0.951, Recall = 0.824), demonstrating strong generalization beyond our study dataset.
Cross-annotator agreement. We randomly sampled 500 tweets from our dataset and obtained independent annotations from two additional annotators using identical classification criteria. Inter-annotator agreement measured by Cohen’s ranged from 0.71 to 0.79 for stance classification and 0.72 to 0.86 for theme classification, indicating substantial agreement [49].
Statistical analysis
This study employs multiple logistic regression models to evaluate the association between Musk’s termination of X’s COVID misinformation policy and changes in anti-vaccine discourse, focusing on changes in the proportion of anti-vaccine tweets and the distribution of anti-vaccine topics. All models are estimated separately for four types of tweets: total population, original tweets, quoted tweets, and retweets.
We construct two sets of logistic regression models to address distinct research questions. The first set examines whether policy termination was associated with a rise in anti-vaccine tweets. In these models, the dependent variable (Y) is a binary indicator, where Y = 1 if the tweet expresses an anti-vaccine stance and Y = 0 if it supports vaccination or remains neutral. The second set of models investigates whether policy termination was associated with changes in the thematic distribution of anti-vaccine tweets. Here, the dependent variable (Y) is also binary, with Y = 1 indicating that the tweet pertains to a specific anti-vaccine topic and Y = 0 meaning the tweet does not address that topic. Since each topic is modeled separately, this results in multiple logistic regression analyses.
The independent variable in all models is policy change (Xpolicy), where Xpolicy = 0 represents the pre-policy termination period (November 16–22, 2022), and Xpolicy = 1 represents the post-policy termination period (November 24–30, 2022). All models report odds ratios, 95% confidence intervals, and p-values, with statistical significance determined at p < 0.05.
Thus, each logistic regression follows the general form shown in Eq (1):
where Y represents the dependent variable, which varies based on the model type as described above. This analysis aims to reveal whether the overall prevalence and thematic structure of anti-vaccine discourse changed following policy termination.
To assess the robustness of our findings, we also conducted sensitivity analyses excluding highly retweeted content (≥ 50 retweets) to examine whether results were driven by a small number of widely shared posts.
Results
Exploratory analysis of identified vaccine stances and topics
The final annotated dataset comprised 13,458 tweets: 5,896 from the pre-policy period (November 16–22) and 7,562 from the post-policy period (November 24–30). By tweet type, the dataset included 2,444 original tweets (18.2%), 10,001 retweets (74.3%), and 1,013 quote tweets (7.5%). Regarding vaccine stance, 6,674 tweets (49.6%) were classified as anti-vaccine, 3,949 (29.3%) as neutral, 2,442 (18.1%) as pro-vaccine, and 393 (2.9%) had no vaccine stance mentioned. This distribution yielded a near-balanced classification between anti-vaccine (49.6%) and non-anti-vaccine content (50.4%).
This subsection presents an exploratory analysis of vaccine stance tweets annotated by GPT-4o, examining trends before and after X terminated its COVID-19 misinformation policy. Fig 1 illustrates the temporal trends of different vaccine stances, highlighting a pronounced increase in tweet volume immediately following the policy change (Nov 23). Notably, this increase was driven largely by anti-vaccine tweets, whereas pro-vaccine and neutral tweets exhibited relatively smaller changes. Fig 2 further decomposes these trends across original tweets (TW), retweets (RT), and quote tweets (QT), showing a particularly marked rise in retweet activity. This indicates that the policy shift mainly amplified existing vaccine discourse rather than prompting new original content creation. Fig 3 reinforces this pattern by quantitatively comparing tweet types, clearly revealing that retweeting behavior was the primary mechanism amplifying vaccine-related narratives post-policy termination.
Each color represents a distinct stance category: anti-vaccine, neutral, pro-vaccine, and no mention of vaccine stance. The dashed line denotes the policy change date (Nov 23). The y-axis shows the percentage of each stance category, highlighting the relative shift in composition following the policy update.
These three subplots illustrate vaccine stance trends across original tweets (TW), retweets (RT), and quote tweets (QT). The dashed line represents the policy change date. The y-axis shows the percentage of each stance category within each tweet type.
Solid bars represent pre-policy data, while hatched bars indicate post-policy data. Percentages are shown above each bar, facilitating direct comparison of compositional shifts across tweet types.
To further contextualize these visual patterns, Tables 6 and 7 provide examples and thematic distributions of non-anti-vaccine and anti-vaccine tweets before and after the policy shift. Among non-anti-vaccine tweets, neutral content increased proportionally, whereas explicitly pro-vaccine tweets showed a modest decrease. Anti-vaccine tweets exhibited notable thematic shifts, with health concerns, vaccine inefficacy, and conspiracy narratives becoming more prominent after the policy termination. Fig 4 shows the overall distribution of anti-vaccine themes, while Fig 5 illustrates how the thematic composition evolved over time before and after the policy change.
The most prevalent theme is health concerns, followed by against mandatory vaccination and deeper conspiracy. Less common topics include shedding and Big Pharma. Percentages are shown above each bar, illustrating the relative prevalence of vaccine hesitancy narratives.
Each color represents a different theme, and the dashed line marks the policy change date (Nov 23). The y-axis shows the percentage composition of topics, revealing how the relative prominence of health concerns, government conspiracies, and vaccine inefficacy narratives shifted following the policy change.
Does policy termination lead to an increase in anti-vaccine tweets?
To statistically evaluate whether X’s termination of the COVID-19 misinformation policy was significantly associated with the prevalence of anti-vaccine tweets, we conducted regression analyses (Fig 6). The results indicate that the policy termination was significantly associated with an increase in anti-vaccine tweets within the total population of analyzed tweets (odds ratio [OR] = 1.60, 95% CI: 1.50–1.72, p < 0.001). Specifically, retweets exhibited the most substantial increase in anti-vaccine sentiment following policy termination (OR = 1.75, 95% CI: 1.62–1.90, p < 0.001), highlighting retweets as the primary mechanism amplifying existing anti-vaccine content. Original tweets also showed a statistically significant, albeit more moderate, increase in anti-vaccine stance (OR = 1.41, 95% CI: 1.19–1.67, p < 0.001). Conversely, quoted tweets did not demonstrate a statistically significant change (OR = 0.98, 95% CI: 0.77–1.26, p = 0.893), indicating no meaningful shift in anti-vaccine stance among quote tweets following the policy termination.
Odds ratios with 95% confidence intervals are presented for different tweet types. The overall population exhibits a significant increase in anti-vaccine tweets, while quoted tweets show no statistically significant change.
To address concerns about highly retweeted content disproportionately influencing results, we conducted a sensitivity analysis by excluding tweets that were retweeted 50 or more times. This filtering removed 30 original tweets and their 5,407 associated retweets (40.2% of the data). Table 8 presents representative examples of highly retweeted anti-vaccine content.
The sensitivity analysis results reveal an important nuance: while the overall effect of policy termination remained significant, the retweet-specific effect was substantially attenuated and no longer statistically significant. Notably, the effect for original tweets increased, suggesting that the policy change was associated with the creation of new anti-vaccine content rather than merely amplifying existing narratives. This finding underscores the importance of distinguishing between content amplification and content generation when assessing policy impacts on online discourse (Fig 7).
Hatched bars indicate tweets from the post-policy period (Nov 24–30). Despite having shorter observation windows for accumulating retweets (0–6 days versus 8–14 days for pre-policy tweets), post-policy tweets comprised 50% of the top 20 most retweeted tweets, suggesting accelerated spread of anti-vaccine content following policy termination.
Does the policy termination significantly shift the prevalence of anti-vaccine themes?
To assess whether X’s termination of its COVID-19 misinformation policy was significantly associated with the thematic prevalence of anti-vaccine tweets, we conducted regression analyses focusing on four predominant anti-vaccine themes: health concerns, deeper conspiracy, opposition to mandatory vaccination, and vaccine ineffectiveness. Fig 8 presents the complete statistical results.
Each panel corresponds to a distinct anti-vaccine narrative, showing whether policy change was associated with its prevalence. Statistically significant increases and decreases are observed for various tweet types.
Our results reveal distinct thematic shifts following policy termination. Tweets emphasizing health concerns significantly increased overall, particularly driven by retweets and quoted tweets, while original tweets showed no statistically significant change. This indicates that the policy shift notably amplified user engagement in disseminating existing health-risk narratives rather than generating new content on this theme.
In contrast, tweets associated with deeper conspiracy narratives and opposition to mandatory vaccination both exhibited significant overall decreases, primarily driven by reduced retweeting activity. Original tweets and quoted tweets for these themes showed no significant change, suggesting that while these narratives were less frequently amplified, their original creation remained stable.
The vaccine ineffectiveness theme presented a more nuanced pattern. While overall prevalence decreased, original tweets on this theme showed a statistically significant increase. This suggests that users were more likely to generate new content questioning vaccine effectiveness post-policy termination, even as the overall retweeting of such content declined.
Impact of policy termination on stance consistency in quote tweets
We further analyzed whether X’s policy termination impacted the consistency of vaccine stances expressed by users in quote tweets relative to the original quoted content (Fig 9). Overall, anti-vaccine quote tweets exhibited the highest stance consistency (255 tweets), while neutral tweets were most frequently classified as unclear (152 tweets), indicating challenges in accurately interpreting neutral user attitudes from their commentary.
The leftmost heatmap represents the overall distribution, while the middle and right heatmaps correspond to the pre-policy and post-policy periods, respectively. The color intensity reflects the frequency of each category, with darker shades indicating higher tweet counts. Notably, anti-vaccine quote tweets show increased alignment with the quoted content after the policy change.
Comparing tweet distributions before and after policy termination reveals notable shifts. Before the policy change, anti-vaccine quote tweets showed nearly equal proportions of consistent (109 tweets) and inconsistent stances (100 tweets), with fewer unclear cases (57 tweets). After policy termination, there was a substantial increase in consistent anti-vaccine stances (146 tweets), accompanied by a marked decrease in inconsistent stances (41 tweets). The unclear cases among anti-vaccine tweets also increased (72 tweets), though less markedly. This pattern suggests a growing tendency among users to reinforce existing anti-vaccine messages after content moderation ceased.
Neutral quote tweets maintained a predominantly unclear stance both before (80 tweets) and after (72 tweets) policy termination, suggesting persistent ambiguity in neutral vaccine discussions. Pro-vaccine quote tweets exhibited relatively stable distributions over time, consistently showing a clear alignment with quoted content.
Discussion
Principal findings
This study examined the changes following X’s termination of its COVID-19 vaccine misinformation policy on vaccine-related discourse, addressing three research questions. For anti-vaccine tweet volume (RQ1), our findings indicate that the proportion of anti-vaccine tweets significantly increased after the policy change, particularly in retweets and quote tweets. Meanwhile, the proportion of pro-vaccine tweets declined, whereas neutral tweets became more prevalent. This suggests that the policy adjustment was associated with changes in the volume of anti-vaccine content but also altered the overall information landscape, where pro-vaccine discourse became relatively less visible. Notably, sensitivity analyses excluding highly retweeted content revealed that the policy change was associated with increased creation of new anti-vaccine content, rather than merely amplifying existing narratives through retweets.
Importantly, our primary analysis intentionally retained retweets because they constitute the information environment to which users are actually exposed. From a public health perspective, it is the prevalence of anti-vaccine content in users’ feeds, regardless of whether it originates from new posts or amplified existing ones, that shapes exposure and potentially influences attitudes [3]. Our sensitivity analysis complements this by distinguishing between content amplification and content generation, addressing whether the observed increase reflects viral spread of a few posts or broader discourse changes. Together, these analyses provide a more complete picture: the main analysis captures changes in the information landscape users encounter, while the sensitivity analysis confirms that the policy change was associated with both amplification of existing content and creation of new anti-vaccine discourse.
For thematic shifts (RQ2), the results show that the thematic distribution of anti-vaccine tweets changed after the policy termination. Specifically, tweets expressing concerns about vaccine safety and health risks became more prominent, while changes in themes related to distrust in pharmaceutical companies and opposition to vaccine mandates were relatively minor. Additionally, some conspiracy-based narratives, such as those linking vaccines to government control or microchip implantation, declined. These findings suggest that after the policy termination, anti-vaccine discourse increasingly centered on vaccine safety concerns rather than adopting more extreme conspiracy narratives.
For stance consistency in quote tweets (RQ3), our analysis found that after the policy termination, anti-vaccine users were more likely to quote tweets aligned with their own stance and less likely to engage with opposing views. This suggests that the policy change not only increased the volume of anti-vaccine tweets but also reinforced ideological polarization, leading to decreased cross-stance interactions in vaccine discourse.
Comparison with prior literature
Our findings align with and extend prior research on the dissemination of vaccine discourse on social media. Previous studies have shown that anti-vaccine narratives often rely on emotionally charged and polarized rhetoric, making them more engaging and widely shared than pro-vaccine content [19,20]. Our study further confirms this trend, showing that anti-vaccine tweets became even more pronounced after the policy change, particularly through retweets [28,29]. This amplification pattern is consistent with Mønsted and Lehmann’s [27] characterization of polarization in online vaccine discourse.
Regarding the impact of content moderation policies, prior research has primarily focused on their implementation and effectiveness in reducing the spread of misinformation. Studies have shown that interventions such as content labeling and fact-checking can reduce engagement with misleading narratives [9,31]. However, fewer studies have examined the short-term effects of policy removal [41]. Our study fills this gap by documenting that the termination of X’s misinformation policy did not lead to a uniform increase in anti-vaccine discourse but rather a selective amplification of specific themes, particularly vaccine safety concerns. This aligns with the findings of Allen et al. [24], who suggested that vaccine-skeptical narratives, rather than outright false claims, play a stronger role in influencing vaccine hesitancy.
Our findings on stance consistency complement [42], who showed that Facebook’s moderation drove anti-vaccine content toward politicization. We observed a different pattern: increased ideological alignment in quote tweets, suggesting that policy removal was associated with changes in user interaction patterns rather than just content volume.
Methodological contributions
A key contribution of this study is demonstrating the feasibility of using large language models for large-scale social media content analysis. We employed GPT-4o with Chain-of-Thought prompting to perform three classification tasks, stance detection, thematic categorization, and stance consistency assessment, on over 13,000 tweets. This approach offers several advantages over traditional methods.
First, compared to manual annotation, LLM-based classification enables analysis at scale while maintaining consistency. Our validation on the CAVES benchmark [48] achieved F1 = 0.883 for anti-vaccine stance detection, with cross-annotator agreement (Cohen’s = 0.71–0.86) comparable to human inter-rater reliability. Second, unlike keyword-based or traditional machine learning approaches that require extensive labeled training data, our prompt-based method can be rapidly adapted to new classification tasks. Third, the Chain-of-Thought prompting provides interpretable reasoning for each classification decision, enhancing transparency.
These methodological advances complement recent work on automated misinformation detection [50] and suggest that validated LLM classifiers can serve as practical tools for researchers studying online discourse dynamics, particularly when events unfold rapidly and timely analysis is needed.
Policy implications
Our findings have implications for platform governance. Within this short pre-post window, the termination of X’s COVID-19 misinformation policy was associated with greater visibility of anti-vaccine content, especially through retweets. The observed thematic shifts also indicate that this policy change was associated with different short-term patterns across narrative types. Increased stance consistency in quote tweets further suggests greater within-stance alignment in vaccine-related discussion.
These findings highlight the complexity of content moderation decisions. In this case, terminating X’s COVID-19 misinformation policy was associated with short-term shifts in the visibility and composition of vaccine-related discourse. At the same time, our observational design does not permit causal inference. The broader question of how platforms should balance free expression and public health remains open.
Limitations and future directions
Despite the insights provided by this study, several limitations should be acknowledged. First, our analysis is based on a short time window covering only seven days before and after the policy termination, which does not allow for assessment of long-term effects. Additionally, pre-policy tweets had more time to accumulate retweets than post-policy tweets, yet post-policy content still comprised 50% of the top 20 most retweeted tweets (Fig 7), suggesting our estimates may be conservative. Future research should extend the observation period to determine whether the observed changes persist over time.
Second, our pre-post observational design precludes causal inference, as the period surrounding the policy termination was characterized by multiple concurrent events. Our analysis captures changes associated with the announcement of policy termination rather than changes in enforcement practices. Relatedly, our regression models use a single binary predictor (pre/post policy) without additional control variables. This parsimonious specification reflects our short 14-day observation window, during which user composition and platform characteristics are unlikely to change systematically. Stratified analyses by tweet type partially address confounding, but potential confounders such as concurrent news events or algorithm changes cannot be adequately captured through tweet-level covariates. Furthermore, this study focuses on tweet content without analyzing user-level behavior, such as which users were more likely to post or amplify anti-vaccine content, or the role of key opinion leaders. Prior research has shown that different user categories, such as government accounts, media outlets, and ordinary users, play distinct roles in epidemic-related information propagation, with media dominating reach while general users propagate information faster [51]. Future studies could integrate user-level analysis to better understand differential responses to policy changes.
Finally, our study does not explore the social network structures underlying the spread of anti-vaccine discourse. Future research could employ social network analysis to examine how policy termination influenced information diffusion pathways and community dynamics.
Conclusion
This study examined changes in vaccine-related discourse following X’s termination of its misinformation policy, focusing on the prevalence, thematic composition, and stance consistency of anti-vaccine tweets. Our findings indicate that the policy change was associated with a significant increase in anti-vaccine content, particularly through retweets, which served as the primary mechanism for amplification. Sensitivity analyses further revealed that the policy change was associated with both the amplification of existing content and the creation of new anti-vaccine discourse. Thematic analysis indicated that while concerns over vaccine safety intensified, deeper conspiracy narratives and opposition to vaccine mandates declined. Additionally, stance consistency in quote tweets strengthened post-policy termination, suggesting a reinforcement of ideological alignment within anti-vaccine discourse.
These findings contribute to the ongoing debate on content moderation by suggesting that the removal of misinformation policies may be associated not only with an increase in the quantity of controversial discourse but also with changes in thematic structure and user engagement patterns. The results highlight the role of platform governance in shaping public health narratives. However, our pre-post observational design and the presence of concurrent events during the study period preclude definitive causal conclusions. Future research should extend the observation period to examine the long-term effects of policy changes and investigate the role of key influencers and network dynamics in shaping discourse evolution.
References
- 1. Skafle I, Nordahl-Hansen A, Quintana DS, Wynn R, Gabarron E. Misinformation About COVID-19 Vaccines on Social Media: Rapid Review. J Med Internet Res. 2022;24(8):e37367. pmid:35816685
- 2. Machingaidze S, Wiysonge CS. Understanding COVID-19 vaccine hesitancy. Nat Med. 2021;27(8):1338–9. pmid:34272500
- 3. Loomba S, de Figueiredo A, Piatek SJ, de Graaf K, Larson HJ. Measuring the impact of COVID-19 vaccine misinformation on vaccination intent in the UK and USA. Nat Hum Behav. 2021;5(3):337–48. pmid:33547453
- 4. Gisondi MA, Barber R, Faust JS, Raja A, Strehlow MC, Westafer LM, et al. A Deadly Infodemic: Social Media and the Power of COVID-19 Misinformation. J Med Internet Res. 2022;24(2):e35552. pmid:35007204
- 5.
Simon F, Howard PN, Nielsen RK. Types, sources, and claims of COVID-19 misinformation. Reuters Institute for the Study of Journalism. 2020.
- 6. Cinelli M, Quattrociocchi W, Galeazzi A, Valensise CM, Brugnoli E, Schmidt AL, et al. The COVID-19 social media infodemic. Scientific reports. 2020;10(1):16598.
- 7. Ferrara E, Cresci S, Luceri L. Misinformation, manipulation, and abuse on social media in the era of COVID-19. J Comput Soc Sci. 2020;3(2):271–7. pmid:33251373
- 8. Poddar S, Mondal M, Misra J, Ganguly N, Ghosh S. Winds of Change: Impact of COVID-19 on Vaccine-Related Opinions of Twitter Users. ICWSM. 2022;16:782–93.
- 9. Kozyreva A, Herzog SM, Lewandowsky S, Hertwig R, Lorenz-Spreen P, Leiser M, et al. Resolving content moderation dilemmas between free speech and harmful misinformation. Proc Natl Acad Sci U S A. 2023;120(7):e2210666120. pmid:36749721
- 10.
Twitter. Updating our approach to misleading information. 2020. https://blog.x.com/en_us/topics/product/2020/updating-our-approach-to-misleading-information
- 11.
BBC News. Coronavirus: Twitter will label Covid-19 fake news. 2020. https://www.bbc.com/news/technology-52632909
- 12.
Meta. An Update on Our Work to Keep People Informed and Limit Misinformation About COVID-19. Meta. 2020. https://about.fb.com/news/2020/04/covid-19-misinfo-update/
- 13.
Instagram Help Center. COVID-19 and Vaccine Policy Updates and Protections. Instagram Help Center. https://help.instagram.com/697825587576762. 2024.
- 14.
YouTube Help. Medical misinformation policy - YouTube Help. 2024. https://support.google.com/youtube/answer/13813322?hl=en.2024
- 15. Krishnan N, Gu J, Tromble R, Abroms LC. Research note: Examining how various social media platforms have responded to COVID-19 misinformation. Harvard Kennedy School Misinformation Review. 2021;2(6):1–25.
- 16.
BBC News. Twitter ends Covid misinformation policy under Musk. 2022. https://www.bbc.com/news/technology-63796832
- 17.
Allyn B. Elon Musk bought Twitter. Here’s what he says he’ll do next. NPR. 2022.
- 18.
Katsuyama J. Elon Musk takes over Twitter, saying it will be “digital town square” not a “free-for-all”. KTVU FOX 2. 2022. https://www.ktvu.com/news/elon-musk-takes-over-twitter-saying-it-will-be-digital-town-square-not-a-free-for-all
- 19. Nguyen A, Catalan-Matamoros D. Anti-Vaccine Discourse on Social Media: An Exploratory Audit of Negative Tweets about Vaccines and Their Posters. Vaccines (Basel). 2022;10(12):2067. pmid:36560477
- 20. Ward JK, Peretti-Watel P, Verger P. Vaccine criticism on the Internet: Propositions for future research. Hum Vaccin Immunother. 2016;12(7):1924–9. pmid:26900646
- 21. Blankenship EB, Goff ME, Yin J, Tse ZTH, Fu K-W, Liang H, et al. Sentiment, Contents, and Retweets: A Study of Two Vaccine-Related Twitter Datasets. Perm J. 2018;22:17–138. pmid:29911966
- 22.
BBC News. Coronavirus: Bill Gates ‘microchip’ conspiracy theory and other vaccine claims fact-checked. BBC News. 2020. https://www.bbc.com/news/52847648
- 23. Dredze M, Broniatowski DA, Smith MC, Hilyard KM. Understanding Vaccine Refusal: Why We Need Social Media Now. Am J Prev Med. 2016;50(4):550–2. pmid:26655067
- 24. Allen J, Watts DJ, Rand DG. Quantifying the impact of misinformation and vaccine-skeptical content on Facebook. Science. 2024;384(6699):eadk3451. pmid:38815040
- 25. Broniatowski DA, Jamison AM, Qi H, AlKulaib L, Chen T, Benton A. COVID-19 misinformation is less likely to spread than is accurate COVID-19 content. PLOS ONE. 2022;17(10):e0275417.
- 26. Schmidt AL, Zollo F, Scala A, Betsch C, Quattrociocchi W. Polarization of the vaccination debate on Facebook. Vaccine. 2018;36(25):3606–12. pmid:29773322
- 27. Mønsted B, Lehmann S. Characterizing polarization in online vaccine discourse-A large-scale study. PLoS One. 2022;17(2):e0263746. pmid:35139121
- 28. Broniatowski DA, Jamison AM, Qi S, AlKulaib L, Chen T, Benton A, et al. Weaponized Health Communication: Twitter Bots and Russian Trolls Amplify the Vaccine Debate. Am J Public Health. 2018;108(10):1378–84. pmid:30138075
- 29. Thelwall M, Kousha K, Thelwall S. Covid-19 vaccine hesitancy on English-language Twitter. Profesional de la Información. 2021;30(2):e300212.
- 30. Hwang H, Shim J, Lee S, Kim S. Vaccine discourse on Twitter during the onset of COVID-19: topic modelling and influence analysis. PLOS ONE. 2022;17(11):e0276634.
- 31. Pennycook G, McPhetres J, Zhang Y, Lu JG, Rand DG. Fighting COVID-19 Misinformation on Social Media: Experimental Evidence for a Scalable Accuracy-Nudge Intervention. Psychol Sci. 2020;31(7):770–80. pmid:32603243
- 32. Bode L, Vraga EK. In Related News, That Was Wrong: The Correction of Misinformation Through Related Stories Functionality in Social Media. J Commun. 2015;65(4):619–38.
- 33. Pennycook G, Cannon TD, Rand DG. Prior exposure increases perceived accuracy of fake news. J Exp Psychol Gen. 2018;147(12):1865–80. pmid:30247057
- 34. Horta Ribeiro M, Jhaver S, Zannettou S, Blackburn J, Stringhini G, De Cristofaro E, et al. Do Platform Migrations Compromise Content Moderation? Evidence from r/The_Donald and r/Incels. Proc ACM Hum-Comput Interact. 2021;5(CSCW2):1–24.
- 35. Johnson NF, Leahy R, Restrepo NJ, Velasquez N, Zheng M, Manrique P, et al. Hidden resilience and adaptive dynamics of the global online hate ecology. Nature. 2019;573(7773):261–5. pmid:31435010
- 36. Starbird K. Disinformation’s spread: bots, trolls and all of us. Nature. 2019;571(7766):449. pmid:31341295
- 37. Cunningham D, Everton S. A network analysis of Twitter’s crackdown on the QAnon conversation. Journal of Social Structure. 2022;23(1):4–27.
- 38.
Aliapoulios M, Bevensee E, Blackburn J, Bradlyn B, De Cristofaro E, Stringhini G. An early look at the parler online social network. 2021. https://doi.org/10.48550/arXiv.2101.03820
- 39.
Juneja P, Bhuiyan MM, Mitra T. Assessing enactment of content regulation policies: A post hoc crowd-sourced audit of election misinformation on YouTube. In: Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, 2023. 1–22. https://doi.org/10.1145/3544548.3580846
- 40. Lanier HD, Diaz MI, Saleh SN, Lehmann CU, Medford RJ. PLOS ONE. 2022;17(6):e0268681.
- 41.
Mitts T, Pisharody N, Shapiro J. Removal of Anti-Vaccine Content Impacts Social Media Discourse. In: 14th ACM Web Science Conference 2022, 2022. 319–26. https://doi.org/10.1145/3501247.3531548
- 42. Broniatowski DA, Simons JR, Gu J, Jamison AM, Abroms LC. The efficacy of Facebook’s vaccine misinformation policies and architecture during the COVID-19 pandemic. Sci Adv. 2023;9(37):eadh2132. pmid:37713497
- 43. Hobbs A, Aldosery A, Kostkova P. Low credibility URL sharing on Twitter during reporting linking rare blood clots with the Oxford/AstraZeneca COVID-19 vaccine. PLoS One. 2024;19(1):e0296444. pmid:38241268
- 44. Banda JM, Tekumalla R, Wang G, Yu J, Liu T, Ding Y, et al. A Large-Scale COVID-19 Twitter Chatter Dataset for Open Scientific Research-An International Collaboration. Epidemiologia (Basel). 2021;2(3):315–24. pmid:36417228
- 45. Pierri F, DeVerna MR, Yang K-C, Axelrod D, Bryden J, Menczer F. One Year of COVID-19 Vaccine Misinformation on Twitter: Longitudinal Study. J Med Internet Res. 2023;25:e42227. pmid:36735835
- 46. Firdaus SN, Ding C, Sadeghian A. Retweet: A popular information diffusion mechanism – A survey paper. Online Social Networks and Media. 2018;6:26–40.
- 47.
Ramage D, Hall D, Nallapati R, Manning CD. Labeled LDA: A supervised topic model for credit attribution in multi-labeled corpora. In: 2009. 248–56.
- 48.
Poddar S, Samad AM, Mukherjee R, Ganguly N, Ghosh S. In: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, 2022. 556–66.
- 49. Landis JR, Koch GG. The measurement of observer agreement for categorical data. Biometrics. 1977;33(1):159–74. pmid:843571
- 50. Poddar S, Mukherjee R, Khatuya S, Ganguly N, Ghosh S. How COVID-19 Has Impacted the Anti-vaccine Discourse: A Large-Scale Twitter Study Spanning Pre-COVID and Post-COVID Era. ICWSM. 2024;18:1276–88.
- 51. Sun L, Pang PC-I, Si Y-W. Roles of information propagation of Chinese microblogging users in epidemics: a crisis management perspective. INTR. 2020;31(2):540–61.