Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Exploring scalable assessment methods for terminated trials in ClinicalTrials.gov: A cohort analysis of German and Californian trials

  • Samruddhi Suresh Yerunkar ,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing

    samruddhi.yerunkar@charite.de

    Affiliations Berlin Institute of Health at Charité – Universitätsmedizin Berlin, QUEST Center for Responsible Research, Charitéplatz 1, Berlin, Germany, Department of Psychiatry and Psychotherapy, Charité - Universitätsmedizin Berlin, Campus Benjamin Franklin, Berlin, Germany

  • Benjamin Gregory Carlisle,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Supervision, Writing – review & editing

    Affiliation STREAM research group, Department of Ethics, Equity and Policy, School of Population and Global Health, McGill University, Montréal, Canada

  • Delwen L. Franzen,

    Roles Conceptualization, Methodology, Writing – review & editing

    Affiliation Berlin Institute of Health at Charité – Universitätsmedizin Berlin, QUEST Center for Responsible Research, Charitéplatz 1, Berlin, Germany

  • Maia Salholz-Hillel,

    Roles Methodology, Writing – review & editing

    Affiliation Berlin Institute of Health at Charité – Universitätsmedizin Berlin, QUEST Center for Responsible Research, Charitéplatz 1, Berlin, Germany

  • Daniel Strech,

    Roles Conceptualization, Methodology, Project administration, Supervision, Writing – review & editing

    Affiliation Berlin Institute of Health at Charité – Universitätsmedizin Berlin, QUEST Center for Responsible Research, Charitéplatz 1, Berlin, Germany

  • Susanne Gabriele Schorr

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Supervision, Validation, Visualization, Writing – review & editing

    Affiliation Berlin Institute of Health at Charité – Universitätsmedizin Berlin, QUEST Center for Responsible Research, Charitéplatz 1, Berlin, Germany

Abstract

Introduction

Clinical trials can terminate early for many reasons, including non-scientific reasons. We aimed to develop scalable semi-automated methods to characterize terminated trials and explore a methodology to estimate the risk of experiencing serious adverse events (SAEs) in trials terminated due to non-scientific reasons.

Methods

Two cohorts of clinical trials registered in ClinicalTrials.gov were investigated: (1) a cohort of clinical trials affiliated with German university medical centers (reported as completed between 2009–2017) and (2) a cohort of clinical trials affiliated with Californian university medical centers (reported as completed between 2014–2017). We used these cohorts to explore scalable assessment methods and compare terminated trials to completed ones regarding trial characteristics, including therapeutic focus. In a subset of trials terminated for non-scientific reasons with tabular summary results and a parallel, randomized design, we estimated additional risk for SAE. For the German cohort, if results were missing from ClinicalTrials.gov, results from the EU Clinical Trials Register (EUCTR) were included if available.

Results

Of 2,253 German and 1,091 Californian trials, 217 (10%) and 150 (14%) were terminated, respectively. The majority (German: 65%, Californian: 67%) cited non-scientific reasons for termination, primarily low accrual. Compared to completed trials, terminated trials showed lower rates of results reporting: 35% vs. 78% in Germany and 72% vs. 87% in California. Of 242 trials terminated for non-scientific reasons, 14 (6%; 11 from ClinicalTrials.gov, 3 from EUCTR) could be included in the SAE risk assessment. In this limited exploratory analysis, no difference in SAE risk was observed between the intervention and control arms (RR 1.05, 95% CI: 0.76–1.44).

Discussion

Results of terminated trials were less frequently reported, limiting opportunities for knowledge generation. Broader adoption of harmonized result reporting standards across registries, structured templates, and improved logic checks could enable scalable assessment approaches and enhance the utility of terminated trial data for clinical research transparency.

Introduction

Clinical trials are the cornerstone of evidence-based healthcare, providing essential information on the safety and efficacy of new treatments and interventions. Rigorously planned and conducted randomized clinical trials are critical for informing healthcare decisions [1], with various guidelines in place to ensure they are conducted with the highest standards of safety, ethics, and integrity. Besides informing health care decisions, results and methods used in clinical trials are also relevant in shaping the design of new studies [2,3]. However, despite their importance and structured planning processes, not all clinical trials can reach their pre-defined goals, and some discontinue prematurely (terminate).

Speich et al. (2022) analyzed 326 trials approved by research ethics committees in Switzerland, the United Kingdom (UK), Germany, and Canada and reported that 30% (98/326) of trials were prematurely discontinued. The study also highlights that prematurely discontinued trials were less likely to have their results published in journals or reported in registries compared to completed trials [1]. Premature discontinuation of trials can occur due to scientific reasons, such as ethical concerns related to efficacy or safety, or non-scientific reasons such as poor project planning or low enrollment rates [4,5]. Williams et al. (2015) highlighted that, as of February 2013, the most common reason for trial termination among terminated trials was non-scientific reasons [5]. This is particularly concerning because this might, in principle, have been foreseen or prevented, leading to unnecessary expenditure of time and resources. Most importantly, participants who invest their time to participate in clinical research might be exposed to interventions and measures with unclear benefits and risk experiencing adverse events.

Clinical trial interventions are often experimental and carry a higher risk of adverse events than standard clinical practice [6]. Serious adverse events (SAEs) are defined by ClinicalTrials.gov as ‘any untoward medical occurrence that at any dose results in death, is life-threatening, requires inpatient hospitalization or prolongation of existing hospitalization, results in persistent or significant disability/incapacity, or is a congenital anomaly/birth defect’ [7]. Studies have consistently shown that reporting of SAEs is more comprehensive on ClinicalTrials.gov compared to reporting of SAEs in published journal articles [810], enabling better safety data synthesis. Assessing SAE data available in such trial registries can offer additional context for understanding safety aspects in clinical research [11]. Given that most premature terminations occur for non-scientific reasons, investigating SAE data from these trials may help to understand the potentially avoidable risk faced by participants enrolled in these trials.

Our group has previously investigated how University Medical Centers (UMC) affiliated clinical trials in Germany and California perform on transparency practices, such as prospective registration and results reporting [3,12,13]. For prematurely discontinued trials in those cohorts, we did not assess the reason for termination or how they differed from completed trials. In this study, we aimed to develop scalable semi-automated methods to characterize terminated trials in these two cohorts, quantify reasons for termination, and explore a methodology to estimate the risk of experiencing SAEs in trials terminated due to non-scientific reasons.

Materials and methods

The protocol for this project was preregistered on the Open Science Framework (OSF) on November 21, 2023, and is available at https://osf.io/n4ujs/. The study is reported according to the Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) guideline for cross-sectional studies [14].

Data sources

We included clinical trials from two previously published cohorts:

  1. (a) The German UMCs cohort includes all interventional clinical trials registered on ClinicalTrials.gov or the German Clinical Trials Register (Deutsches Register Klinischer Studien, DRKS), affiliated with a German UMC and reported as complete between 2009 and 2017. German UMC affiliation was determined based on the UMC being listed as responsible party, lead sponsor, or having a principal investigator (PI) from the UMC. This dataset is openly available on GitHub (https://github.com/maia-sh/intovalue-data) [15].
  2. (b) The Californian UMCs cohort includes all interventional clinical trials registered on ClinicalTrials.gov, affiliated with a Californian UMC and reported as complete between 2014 and 2017. The dataset is openly available on GitHub (https://github.com/ontogenerator/california-clinical-dashboard) [16].

For both cohorts, publications reporting trial results were identified manually, either through links in the registry or, if no links were available, via Google searches using the trial ID or other matching parameters. Detailed search methodology can be found in the original cohort studies [3,12,13]. These datasets were selected due to their public availability and prior manual identification of results publication.

Eligibility criteria

For both cohorts, inclusion criteria for this study were defined as follows:

Registry: Only trials registered on ClinicalTrials.gov.

Recruitment status: Trials with recruitment status of ‘terminated’ in the ClinicalTrials.gov historical version dated December 1, 2023 were eligible. This version was downloaded using the cthist R package [17]. Terminated trials are defined by ClinicalTrials.gov as studies in which recruitment or enrollment of participants was halted prematurely, participants were no longer being examined or receiving intervention, and the trial would not resume. Trials with a status of completed, defined as studies that concluded as planned with all participants having completed their final visits, were included for comparison [18].

Participant: At least one participant enrolled.

Intervention: All intervention types were included.

Trial study phase: All phases were included.

Data extraction and analyses

Characteristics of the trials (phase, manually identified results publications) were taken from the original datasets. To generate version-specific variables for assessing terminated trials, we developed a custom R package named “terminated_trials_study”, which is openly available on GitHub under the GNU Affero General Public License v3.0 (AGPL-3.0). This package uses the cthist R package [17] to download clinical trial registry entry histories in a structured format. Based on this data, we generated the following variables: degree of enrollment, trial days, summary results and reason for termination. In addition, we developed a novel R package, TrialFociMapper [19], to assign therapeutic foci for trials registered in ClinicalTrials.gov.

‘Degree of enrollment’ was defined by the ratio of actual and anticipated enrollment, expressed as a percentage of enrollment when the trial is terminated. In ClinicalTrials.gov, the enrollment variable is ‘estimated’ at the start of the trial and can be updated to ‘actual’ at various stages of the trial, including at its completion/termination. The function defines anticipated enrollment as the estimated enrollment reported in the first version on or after the trial’s start date and actual enrollment as actual enrollment recorded at the end of the trial. If either anticipated or actual was missing, the function flagged a warning about the missing data. For records where an anticipated number was specified but not explicitly marked as anticipated, we manually assigned it as anticipated. Other records with unclear or missing enrollment information were excluded. For details see degree of enrolment in S1 File.

‘Trial days’ were defined by the number of days between the trial’s stop and start dates. The stop date is defined as the primary completion date reported when its overall status was first changed to terminated from any other status. The ‘start date’ is obtained from the same version of the trial record as the stop date.

The availability of summary results was assessed using the final version of the trial record on ClinicalTrials.gov. Trials investigating investigational medicinal products in Germany are legally required to be registered and to report summary results in the EU Clinical Trials Register (EUCTR). For this German cohort, we additionally applied a custom scraper function to extract summary results from EUCTR. This allowed us to capture additional reporting that may not have been available on ClinicalTrials.gov. Summary results were considered available if present in either ClinicalTrials.gov or the EUCTR registry.

We automatically retrieved the reasons for termination from the ClinicalTrials.gov registry entry ‘why_stopped’ field. This field contains free-text explanations (limited to 160 characters). Two authors (SGS and SSY) manually categorized reasons for termination using a categorization table developed based on existing studies and external sources [1,4,5,2023]. If both non-scientific and scientific reasons were mentioned, the scientific reason was considered primary. In cases where multiple scientific or multiple non-scientific reasons were provided, the description and sequence of reasons were used as a reference (see S1 File, Reason for termination categorization). Discrepancies between authors were resolved by discussion, and if no agreement was reached, a third author (BGC) was consulted. Inter-rater agreement was calculated for the agreement between the two raters (SGS and SSY).

The TrialFociMapper package retrieves the ‘browse_conditions’ table from the AACT (Aggregate Analysis of ClinicalTrials.gov) [24] database, which contains Medical Subject Headings (MeSH) terms submitted by data submitters. These MeSH terms, developed by the National Library of Medicine (NLM), are mapped to high-level categories based on the hierarchy of the MeSH tree. For instance, if a data submitter provides “breast cyst” as a MeSH term, the function assesses the MeSH tree to map “breast cyst” to “neoplasms”, under which it is categorized in the MeSH tree. A trial can have no focus (if not submitted by the data submitter), one focus, or multiple foci. We reported how often each of the therapeutic foci appears across all trials, allowing for multiple foci per study.

Serious adverse events (SAE)

We estimated the risk of experiencing SAEs in trials terminated for non-scientific reasons. This sub-analysis was exploratory in nature, and exclusion criteria had to be adapted during the course of the analysis. We included trials terminated for non-scientific reasons that reported summary results in ClinicalTrials.gov or EUCTR (German cohort). Summary results were only included if they provided data on SAEs in a tabular format. In both ClinicalTrials.gov and EUCTR, adverse event data are structured into separate sections for adverse events and SAEs. Despite some publications providing adverse event data, the inconsistency of reporting between registries and publications [25] prompted us to use only registry-based data for our analysis. We chose to focus on SAEs for this analysis because they are severe, clearly defined, and if summary results are reported, generally well-documented [9]. To estimate the additional risk associated with the intervention, we compared the risk of SAEs in the intervention arm with the risk in the control arm. For this analysis, we included trials with at least two arms that could be clearly assigned to ‘control’ and ‘intervention’. Participants had to be randomized, and at least one participant had received an intervention and thus was at risk of experiencing an SAE. We excluded trials where intervention and control arms could not be assigned, such as single-arm trials, cross-over trials, or trials involving active comparators. We automatically extracted the number of patients at risk and the number of patients experiencing SAEs. We manually categorized the arms into control and intervention to calculate risk ratios and risk differences. For more details on the selection of eligible trials and the manual assignment of arms into control and intervention, see S1 file (SAE risk analysis methodology).

Software

We used the cthist R package [17] to download the historical versions of trials from ClinicalTrials.gov. We developed the terminated-trial-analysis R package [26] to provide functions for this analysis. The R scripts and data generated for this study are available in the terminated-trials-analysis-study repository. Numbat Metanalysis Extraction Manager was utilized for data categorization in our study [27]. The therapeutic focus of each clinical trial was assigned using the R TrialFociMapper package [19]. Data cleaning steps and statistical analyses were performed using R [Version 4.3.2] [28].

Results

General trial characteristics

Out of 2,253 ClinicalTrials.gov registered trials in the German cohort, 1,670 (74%) were completed, and 217 (10%) were terminated. The remaining trials had other recruitment statuses (e.g., withdrawn, unknown) and were not included in the analysis. The 217 terminated trials had planned to enroll 37,379 participants but ultimately enrolled 14,723, resulting in a degree of enrollment of 39% at the time of termination. Among these terminated trials, 18 had enrolled more patients than planned. Completed trials had a degree of enrollment of 96%. Among terminated trials, 29% (n = 64) reported summary results in ClinicalTrials.gov or EUCTR, 32% (n = 69) as publications, and 35% (n = 75) through either format. In comparison, 78% (n = 1306) of completed trials reported results in any format. The median duration until termination was 1,026 days. In terms of therapeutic focus distribution among the terminated trials in the German cohort, neoplasms accounted for 14% of the distribution (n = 60), followed by cardiovascular diseases at 10% (n = 44), nervous system diseases at 7% (n = 32), and infections at 2% (n = 8). See Table 1 (Characteristics of terminated trials and completed trials) for details.

thumbnail
Table 1. Characteristics of terminated trials and completed trials.

https://doi.org/10.1371/journal.pone.0346608.t001

Of the 1,091 trials in the Californian cohort with varied status, 921 (84%) were completed and 150 (14%) terminated. Trials with other recruitment statuses were excluded from the analysis. The enrollment percentage for terminated trials was 37%, compared to 83% for completed trials. The median duration until termination in the Californian cohort was 822 days. Among terminated trials, 62% (n = 93) reported summary results in ClinicalTrials.gov, 32% (n = 48) as publications, and 72% (n = 108) through either format. In comparison, 87% (n = 798) of completed trials reported results in any format. In terms of therapeutic focus distribution among the terminated trials in this cohort, neoplasms accounted for 23% (n = 72) of the distribution, nervous system diseases represented 6% (n = 17), cardiovascular diseases with 4% (n = 12), and infections with 4% (n = 11). For a complete breakdown of all therapeutic foci, see S1 File (Detailed therapeutic foci table).

Reasons for trial termination

Among the 367 terminated trials across both datasets (Germany: 217, California: 150), 241 trials (66%) were terminated for non-scientific reasons, 79 trials (22%) for scientific reasons, and 37 trials (10%) did not provide a reason. Additionally, for 10 (3%) trials, incomplete or vague information made classification difficult (e.g., NCT01439958: ‘Core study 12011.201 was terminated’ and NCT01868503: ‘Protocol modification’). In these cases, raters were unable to determine the exact reason for termination and classified the trials into the “other” category. Overall, the reporting of reasons for trial termination was 96% complete in the Californian cohort and 86% complete in the German cohort. See Table 2 (Reason for trial termination categorization) for details.

thumbnail
Table 2. Reason for trial termination categorization.

https://doi.org/10.1371/journal.pone.0346608.t002

For nine trials reporting both scientific and non-scientific reasons, the scientific reason was prioritized. The interrater reliability (Cohen’s kappa) was 83%. The most common non-scientific reason for termination was ‘low accrual rates’ (Germany: 53%,114; California: 42%,63), whereas the most frequently reported scientific reason was ‘evidence of futility’ (Germany: 9%, 20; California: 8%, 12).

In terms of enrollment, trials terminated for non-scientific reasons had an average enrollment of 34 participants (median: 14) and an average duration of 1092 days (median: 974), while those terminated for scientific reasons enrolled 127 participants (median: 25) with a duration of 1,001 days (median: 885). For further details, see S1 File (Trial characteristics by reason for termination).

SAE risk sub-analysis

Out of 241 trials terminated for non-scientific reasons, 76 (30%) had summary results reported in tabular format in EUCTR or ClinicalTrials.gov. Of these, 61 trials were excluded from the sub-analysis due to design or reporting limitations: 30 were single-arm trials, 20 had active comparators, 7 were non-randomized, 2 employed a crossover design, and 2 had not administered an intervention (see PRISMA flow diagram in S1 File). During iterative development of the SAE analysis methodology, 1 cross-registered trial (NCT01638585, 2010-023426-20) with summary results reported in tabular format in the EUCTR registry; however, these results were provided exclusively as a PDF document rather than in a structured tabular format. Because the comparability of PDF-based results with structured tabular results reported in the EUCTR registry could not be ensured, this trial was excluded in final step.

A total of 14 trials (ClinicalTrials.gov: 11, EUCTR: 3), in which a direct comparison of SAE risk between intervention and control arms was possible, were included in the sub-analysis. We found no statistically significant difference in experiencing SAE between patients in the control arm and patients in the intervention arm (see Fig 1). The common effect model showed a risk ratio (RR) of 1.0674 (95% CI: [0.8625; 1.3210], p = 0.5484), and the random effects model had an RR of 1.0459 (95% CI: [0.7599; 1.4396], p = 0.7830), with low-to-moderate heterogeneity (I² = 27.4%).

thumbnail
Fig 1. Comparison of serious adverse event risk between intervention and control groups in trials terminated for non-scientific reasons.

https://doi.org/10.1371/journal.pone.0346608.g001

Discussion

We developed scalable, semi-automated methods to investigate terminated trials registered in ClinicalTrials.gov, focusing on cohorts of clinical trials affiliated with German and Californian UMCs. Our analysis confirmed that terminated trials had lower rates of results reporting via any route (SR or publication) compared to completed trials, as seen in previous studies [1,5]. This gap reflects barriers to publishing results from terminated trials rather than limited registry reporting. Registries may serve as a valuable platform for disseminating findings from trials that are less likely to appear in journals. Across both datasets, over 60% of trials were terminated for non-scientific reasons, primarily due to low accrual rates. This aligns with previous studies emphasizing patient recruitment as the single main reason for early termination [5,29].

Manual categorization of reasons for trial termination was relatively straightforward for the majority of trials and could potentially be automated. However, a subset posed challenges due to vague or insufficient termination descriptions. In these cases, the authors were unable to determine the exact reason for termination. Implementing predefined categories with an additional free-text option could enhance the analysis of trial terminations in registries. Registry data submitters and authors should also be encouraged to provide clearer explanations for termination, as this data can offer valuable insights for future trials. Since trial registries have limited space for specifying reasons, alternative ways to document key insights should be explored. Brageso et al. (2021) in publishing their findings on the terminated trial NCT03116165, highlighted lessons learned from their work, which were particularly relevant for future work in preventive psychological interventions for post-traumatic stress disorder in emergency settings [30].

We explored a methodology with the potential to be scaled up to assess the SAE risk using registry data for trials terminated due to non-scientific reasons. However, only 6% (14/242) of such trials could be included, as they reported results in the required tabular format on ClinicalTrials.gov or the EUCTR and had comparable intervention and control arms. Identifying eligible trials required extensive manual review, as automated categorization of trial arms was not feasible due to inconsistent labelling of trial arms. For instance, an identifier like ‘EG000’ in ClinicalTrials.gov is typically assumed to represent the experimental arm, but this assumption is incorrect in approximately 25% of cases, as noted in the description of AACT schema, which provides information on data elements in the registry [24]. As a result, manual validation is required to ensure the accurate classification of trial arms.

The primary reason for the low inclusion rate was the limited availability of results reported in the structured tabular format required for our analysis (165/241). Many EUCTR records were excluded because results were submitted in non-tabular formats. While ClinicalTrials.gov mandates that sponsors of terminated trials with enrolled participants submit results including SAEs in a structured, tabular format, EUCTR permits the submission of PDF documents with varying content such as synopses, termination statements, statistical reports, or links. These inconsistencies hinder large-scale analyses and underscore the need for harmonized reporting standards.

Although the Consolidated Standards of Reporting Trials (CONSORT) guidelines [31] have improved reporting quality in academic publications, additional structured guidance is needed for clinical trial registries. A review by Chan et al. (2025) of 16 WHO primary registries including ClinicalTrials.gov and EUCTR revealed substantial variation in how summary results are presented [32]. The new WHO guidance (2025) addresses this by outlining eight minimum elements, including results of benefits and harms, that should be reported in registries [33]. Broader adoption of these standards could improve data consistency and facilitate the use of registry data for analytic purposes, including SAE assessments. Clearly defined reporting templates tailored to different study designs and interventions, together with strengthened quality control and internal logic checks, would further support both within-registry and cross-registry analyses, enhancing the feasibility of applying scalable methodological approaches for assessing SAE risks.

Our exploratory SAE analysis revealed no difference in the risk between the intervention and the control group. However, the limited sample size of our sample can lead to uncertainty, and the findings should therefore be interpreted with caution. Previous studies have employed various classification approaches, such as grouping trials by intervention type or phase, which can lead to different results. For example, an analysis of lifestyle clinical trials (2000–2023) using data from ClinicalTrials.gov found no significant increase in the relative risk of experiencing an SAE in the intervention group compared to the control group [11]. Another study examining phase 3 randomized controlled trials involving solid cancer patients comparing sorafenib with control found a significantly increased risk of SAEs in the intervention group compared to the control group (SAEs: RR 1.49, 95% CI: 1.18–1.89, p < 0.001) [30].

The methods we developed have their advantages. Our approach to characterize trials used historical versions of trial records to capture key dates, such as the primary completion date when trials were first marked as terminated. This approach provides insights into when termination decisions were initially recorded, offering a more updated timeline. Furthermore, our methods for assessing trial duration and enrollment percentage are scalable and can be applied to completed trials. Integrating these metrics with descriptive mappings of therapeutic areas and trial locations can contribute to exploratory analyses. While such data alone cannot determine trial feasibility or predict success, they may support preliminary assessments by identifying areas where early discontinuation has occurred and might indicate potential learnings from terminated trials.

While the methods were designed to maximize reproducibility and scalability, certain limitations, particularly those related to registry data quality and reporting inconsistencies, should be acknowledged. For the German and Californian cohorts, variables such as the availability of trial results publications were derived from the original datasets. As such, the possibility that some of these older trials have published results cannot be ruled out. The two cohorts were used to test the methodology and were drawn from different calendar periods based on data availability. To avoid potential bias, we ensured that all primary analyses were conducted within cohorts only, comparing terminated and completed trials from the same cohort and time period, and no direct comparisons were made between cohorts. We caution readers against drawing direct comparisons between cohorts.

While the assignment of reasons for trial termination and trial arms involved some degree of judgment, we implemented safeguards, including dual independent rating and resolution of disagreements via discussion. Nevertheless, some degree of misclassification cannot be ruled out. In addition, reasons for trial termination were annotated solely based on information provided in the registries. Misclassification may thus arise from incomplete or ambiguous entries or from how information is recorded due to word limits. For example, “slow accrual” may reflect both business-related factors (e.g., insufficient financial incentives) and scientific considerations, such as early signals of limited efficacy or unacceptable toxicity. As a result, business and scientific reasons for termination may overlap and cannot be systematically disentangled using registry data alone.

Therapeutic focus classification was performed using our TrialFociMapper tool, which retrieves and maps focus areas for trials registered in ClinicalTrials.gov. In some cases, a trial may be associated with multiple therapeutic foci. While users can manually review and annotate the most relevant focus, the tool does not yet perform automatic assignment of the most relevant therapeutic focus. In contrast, Brewster et al. (2022) manually reviewed study titles, abstracts, and descriptions for all included trials and assigned each to one of 29 predefined therapeutic foci [4]. While our approach enables automated mapping, we acknowledge that assigning a single, most relevant focus remains an important step. Future development of the tool aims to incorporate an automated selection of the most relevant therapeutic focus per trial. Finally, for the SAE analysis, we relied solely on SAE data reported in trial registries. If SAE information was missing or incomplete in the registry, we did not supplement it with data from journal publications, due to their different reporting formats (tables, text, footnotes) and inconsistencies in SAE counts and definitions between registries and publications [10,34]. We also did not search for additional information outside the trial registry to determine whether SAEs were treatment-related. Finally, in terms of generalizability and scalability, we acknowledge that differences between ClinicalTrials.gov and other trial registries in terms of the types of trials included, the structure of recorded information, and reporting formats may limit the generalizability of our approach.

In conclusion, our findings highlight that terminated trials report results less frequently than completed trials. Yet in a learning research system, data from ‘failed’ or terminated trials can provide valuable insights to inform future studies and help reduce inefficiencies in the clinical research system. Inconsistent reporting requirements hinder analysis, but broader adoption of harmonized standards across registries, structured templates, and registry logic checks could enable scalable assessments and enhance the transparency and utility of terminated trial data.

Supporting information

S1 File. Supporting information file: Includes details about Description: Degree of enrolment - Handling of missing enrollment variable. Table. Reason for termination - categorization. Description: SAE risk analysis methodology and PRISMA flow diagram. Table. Detailed therapeutic foci table5 Table. Trial characteristics by reason for termination.

https://doi.org/10.1371/journal.pone.0346608.s001

(DOCX)

References

  1. 1. Speich B, Gryaznov D, Busse JW, Gloy VL, Lohner S, Klatte K, et al. Nonregistration, discontinuation, and nonpublication of randomized trials: A repeated metaresearch analysis. PLoS Med. 2022;19(4):e1003980. pmid:35476675
  2. 2. Mattina J, Carlisle B, Hachem Y, Fergusson D, Kimmelman J. Inefficiencies and Patient Burdens in the Development of the Targeted Cancer Drug Sorafenib: A Systematic Review. PLoS Biol. 2017;15(2):e2000487. pmid:28158308
  3. 3. Wieschowski S, Riedel N, Wollmann K, Kahrass H, Müller-Ohlraun S, Schürmann C, et al. Result dissemination from clinical trials conducted at German university medical centers was delayed and incomplete. J Clin Epidemiol. 2019;115:37–45. pmid:31195110
  4. 4. Brewster R, Wong M, Magnani CJ, Gunningham H, Hoffer M, Showalter S, et al. Early Discontinuation, Results Reporting, and Publication of Pediatric Clinical Trials. Pediatrics. 2022;149(4):e2021052557. pmid:35314864
  5. 5. Williams RJ, Tse T, DiPiazza K, Zarin DA. Terminated Trials in the ClinicalTrials.gov Results Database: Evaluation of Availability of Primary Outcome Data and Reasons for Termination. PLoS One. 2015;10(5):e0127242. pmid:26011295
  6. 6. Luo J, Eldredge C, Cho CC, Cisler RA. Population Analysis of Adverse Events in Different Age Groups Using Big Clinical Trials Data. JMIR Med Inform. 2016;4(4):e30. pmid:27751983
  7. 7. U.S. Department of Health and Human Services. Common terminology criteria for adverse events (CTCAE) version 4.0. 2009.
  8. 8. Riveros C, Dechartres A, Perrodeau E, Haneef R, Boutron I, Ravaud P. Timing and completeness of trial results posted at ClinicalTrials.gov and published in journals. PLoS Med. 2013;10(12):e1001566; discussion e1001566. pmid:24311990
  9. 9. Chen KY, Borglund EM, Postema EC, Dunn AG, Bourgeois FT. Reporting of clinical trial safety results in ClinicalTrials.gov for FDA-approved drugs: A cross-sectional analysis. Clin Trials. 2022;19(4):442–51. pmid:35482320
  10. 10. Tang E, Ravaud P, Riveros C, Perrodeau E, Dechartres A. Comparison of serious adverse events posted at ClinicalTrials.gov and published in corresponding journal articles. BMC Med. 2015;13:189. pmid:26269118
  11. 11. Key MN, Shaw AR, Erickson KI, Burns JM, Vidoni ED. A retrospective analysis of serious adverse events and deaths in U.S.-based lifestyle clinical trials for cognitive health. Contemp Clin Trials Commun. 2024;38:101277. pmid:38404652
  12. 12. Malički MVN, Wieschowski S, Hildebrand N, Gestrich S, Yerunkar S, Zavalis E, et al. Registration and Reporting of Clinical Trials Affiliated with California Universities and with Primary Completion Date from 2014 to 2017. medRxiv. 2025.
  13. 13. Riedel N, Wieschowski S, Bruckner T, Holst MR, Kahrass H, Nury E. Results dissemination from completed clinical trials conducted at German university medical centers remained delayed and incomplete. The 2014 -2017 cohort. J Clin Epidemiol. 2022;144:1–7. pmid:34906673
  14. 14. von Elm E, Altman DG, Egger M, Pocock SJ, Gøtzsche PC, Vandenbroucke JP, et al. The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) Statement: guidelines for reporting observational studies. Int J Surg. 2014;12(12):1495–9. pmid:25046131
  15. 15. Salholz-Hillel M. Intovalue-data GitHub. https://github.com/maia-sh/intovalue-data
  16. 16. Nachev V. Responsible Metrics dashboard for clinical research transparency GitHub. https://github.com/ontogenerator/california-clinical-dashboard
  17. 17. Carlisle BG. Analysis of clinical trial registry entry histories using the novel R package cthist. PLoS One. 2022;17(7):e0270909. pmid:35776915
  18. 18. ClinicalTrials.gov. Protocol Registration Data Element Definitions for Interventional and Observational Studies. https://clinicaltrials.gov/policy/protocol-definitions
  19. 19. Samruddhi Y, Carlisle BG, Holst MR, Susanne S. Trialfocimapper R package. https://github.com/sama9767/TrialFociMapper 2023.
  20. 20. Carlisle BG, Doussau A, Kimmelman J. Patient burden and clinical advances associated with postapproval monotherapy cancer drug trials: a retrospective cohort study. BMJ Open. 2020;10(2):e034306. pmid:32071183
  21. 21. Deichmann RE, Krousel-Wood M, Breault J. Bioethics in Practice: Considerations for Stopping a Clinical Trial Early. Ochsner J. 2016;16(3):197–8. pmid:27660563
  22. 22. Demetriades AK, Park JJ, Tiefenbach J. Is there resource wastage in the research for spinal diseases? An observational analysis of discontinuation and non-publication in randomised controlled trials. Brain Spine. 2022;2:100922. pmid:36248143
  23. 23. Guideline for good clinical practice E6(R2). 2016.
  24. 24. Clinical Trials Transformation Initiative. Aggregate Analysis of ClinicalTrials.gov (AACT) Database. Durham, NC: Clinical Trials Transformation Initiative. https://aact.ctti-clinicaltrials.org/
  25. 25. Hartung DM, Zarin DA, Guise J-M, McDonagh M, Paynter R, Helfand M. Reporting discrepancies between the ClinicalTrials.gov results database and peer-reviewed publications. Ann Intern Med. 2014;160(7):477–83. pmid:24687070
  26. 26. Samruddhi Y, Carlisle BG. Terminated-trials-study R package. https://github.com/sama9767/terminated-trials-study/tree/main 2023.
  27. 27. Numbat meta-analysis extraction manager. http://bgcarlisle.github.io/Numbat/
  28. 28. R Core Team. R: A language and environment for statistical computing. Vienna, Austria: R Foundation for Statistical Computing. 2022.
  29. 29. Khan MS, Shahid I, Asad N, Greene SJ, Khan SU, Doukky R, et al. Discontinuation and non-publication of heart failure randomized controlled trials: a call to publish all trial results. ESC Heart Fail. 2021;8(1):16–25. pmid:33191637
  30. 30. Bragesjö M, Arnberg FK, Andersson E. Prevention of post-traumatic stress disorder: Lessons learned from a terminated RCT of prolonged exposure. PLoS One. 2021;16(5):e0251898. pmid:34029328
  31. 31. Moher D, Hopewell S, Schulz KF, Montori V, Gøtzsche PC, Devereaux PJ, et al. CONSORT 2010 explanation and elaboration: updated guidelines for reporting parallel group randomised trials. Int J Surg. 2012;10(1):28–55. pmid:22036893
  32. 32. Chan A-W, Karam G, Pymento J, Askie LM, da Silva LR, Aymé S, et al. Reporting summary results in clinical trial registries: updated guidance from WHO. Lancet Glob Health. 2025;13(4):e759–68. pmid:40155113
  33. 33. World Health Organization. Reporting summary results in clinical trial registries: updated guidance from WHO. Geneva: World Health Organization. 2022. https://www.who.int/publications/i/item/9789240065395
  34. 34. Pitrou I, Boutron I, Ahmad N, Ravaud P. Reporting of safety results in published reports of randomized controlled trials. Arch Intern Med. 2009;169(19):1756–61. pmid:19858432