Skip to main content
Advertisement
  • Loading metrics

Biomedical researchers’ perspectives on the reproducibility of research

  • Kelly D. Cobey ,

    Roles Data curation, Formal analysis, Investigation, Methodology, Project administration, Supervision, Validation, Writing – original draft, Writing – review & editing

    kcobey@ottawaheart.ca

    Affiliations University of Ottawa Heart Institute, Ottawa, Canada, School of Epidemiology and Public Health, University of Ottawa, Ottawa, Canada

  • Sanam Ebrahimzadeh,

    Roles Conceptualization, Data curation, Investigation, Methodology, Validation, Writing – review & editing

    Affiliation Centre for Journalology, Ottawa Hospital Research Institute, Ottawa, Canada

  • Matthew J. Page,

    Roles Conceptualization, Investigation, Methodology, Writing – review & editing

    Affiliation Methods in Evidence Synthesis Unit, School of Public Health and Preventive Medicine, Monash University, Melbourne, Australia

  • Robert T. Thibault,

    Roles Investigation, Methodology, Writing – review & editing

    Affiliations Meta-Research Innovation Center at Stanford (METRICS), Stanford University, Stanford, California, United States of America, Coalition for Aligning Science, Chevy Chase, Maryland, United States of America

  • Phi-Yen Nguyen,

    Roles Conceptualization, Data curation, Formal analysis, Methodology, Project administration, Writing – review & editing

    Affiliation Methods in Evidence Synthesis Unit, School of Public Health and Preventive Medicine, Monash University, Melbourne, Australia

  • Farah Abu-Dalfa,

    Roles Formal analysis, Writing – review & editing

    Affiliations University of Ottawa Heart Institute, Ottawa, Canada, School of Political Studies, Faculty of Social Sciences, University of Ottawa, Ottawa, Canada

  • David Moher

    Roles Conceptualization, Investigation, Methodology, Writing – review & editing

    Affiliations School of Epidemiology and Public Health, University of Ottawa, Ottawa, Canada, Centre for Journalology, Ottawa Hospital Research Institute, Ottawa, Canada

Abstract

We conducted an international cross-sectional survey of biomedical researchers’ perspectives on the reproducibility of research. This study builds on a widely cited 2016 survey on reproducibility and provides a biomedical-specific and contemporary perspective on reproducibility. To sample the community, we randomly selected 400 journals indexed in MEDLINE, from which we extracted the author names and emails from all articles published between October 1, 2020 and October 1, 2021. We invited participants to complete an anonymous online survey which collected basic demographic information, perceptions about a reproducibility crisis, perceived causes of irreproducibility of research results, experience conducting reproducibility studies, and knowledge of funding and training for research on reproducibility. A total of 1,924 participants accessed our survey, of which 1,630 provided useable responses (response rate 7% of 23,234). Key findings include that 72% of participants agreed there was a reproducibility crisis in biomedicine, with 27% of participants indicating the crisis was “significant.” The leading perceived cause of irreproducibility was a “pressure to publish” with 62% of participants indicating it “always” or “very often” contributes. About half of the participants (54%) had run a replication of their own previously published study while slightly more (57%) had run a replication of another researcher’s study. Just 16% of participants indicated their institution had established procedures to enhance the reproducibility of biomedical research and 67% felt their institution valued new research over replication studies. Participants also reported few opportunities to obtain funding to attempt to reproduce a study and 83% perceived it would be harder to do so than to get funding to do a novel study. Our results may be used to guide training and interventions to improve research reproducibility and to monitor rates of reproducibility over time. The findings are also relevant to policy makers and academic leadership looking to create incentives and research cultures that support reproducibility and value research quality.

Introduction

There is growing interest in both the reproducibility of research and ways to enhance research transparency [14]. Terminology around reproducibility varies [5]; here, we define reproducibility as re-doing a study using similar methods and obtaining findings consistent with the original study and as irreproducible when the findings are not consistent with the original study. This definition allows for variation in methods (e.g., conceptual and direct replications) between the original study and the reproducibility study as well as different definitions of how “consistent results” are defined (i.e., using p-value, observing results in the same direction, comparing effect sizes). Reproducibility of research is core to maintaining research trustworthiness and fostering translation and progressive discovery. Despite the seemingly critical role of reproducibility of research and growing discussions surrounding reproducibility, the reality is that most studies, including pivotal studies within several disciplines, have never been formally subjected to a reproducibility effort. For example, in education research, an analysis of publications in the field’s top 100 journals showed that just 0.13% publications (221 out of 164,589) described reproducibility projects [6]. In psychology, a study examining 250 articles published between 2014 and 2017 found that 5% described a reproducibility effort [7], while a similar study examining reproducibility in social sciences found that just 1% of articles sampled described a focused on reproducing previous results [8]. Across disciplines, our knowledge about the proportion of studies that are reproducible tends to be dominated by a small number of large-scale reproducibility projects. In psychology, for example, a study which estimated the replicability of 100 foundational studies in top journals in the field reported that only 36% had statistically significant results (one measure of reproducibility), compared to 97% of the original studies [9].

In 2016, Nature reported on a survey of more than 1,500 researchers about their perceptions of reproducibility. They found that 83% agreed there was a reproducibility crisis in science, with 52% indicating that they felt the crisis was “significant” [10]. Survey studies like this play a powerful role in elucidating the determinants of reproducibility. Such information is essential to identify gaps in factors including training, research support, and incentives to ensure reproducible research. Given the global nature of research, capturing global perspectives as was done in the Nature survey is crucial to obtaining broad understanding of issues across the research ecosystem.

In this study, we aim to build on the 2016 Nature survey on reproducibility by surveying researchers in the biomedical community specifically. There is immediate importance to ensuring biomedical research is reproducible: here, studies that were subsequently not reproducible have led to patient harms [11,12]. By capturing a diverse and global group of biomedical researchers’ perceptions of reproducibility within the field we hope to better understand how to ensure reproducibility in biomedicine. While our work is inspired by the 2016 Nature survey, this is not a direct effort to reproduce that study: we sample a different community and use a sampling approach that differs from the original study which limits the ability for direct comparison. Our specific objectives were to: (1) explore biomedical researchers’ perceptions of reproducibility and their perceptions of causes of irreproducibility; and (2) describe biomedical researchers’ experiences conducting and publishing reproducibility projects. The study is descriptive, so we had no formal hypotheses. Understanding researcher perceptions is a key starting point to drive culture change and to create training interventions to drive improvement.

Methods

Open science statement

This study received ethics approval from the Ottawa Health Sciences Research Ethics Board (20210856-01H). This study protocol was posted a priori, and data and materials have been made available: https://osf.io/3ksvz/ [13].

Study design

We conducted an online cross-sectional closed survey of researchers who published a paper in a journal indexed in MEDLINE (RRID:SCR_002185). The survey was anonymous.

Sampling framework

We downloaded the MEDLINE database of journals. From this list of approximately 30,000 journals, we selected a random sample of 400 journals using the RAND() function in Excel (RRID:SCR_016137). We then extracted author names and emails from all articles published in those journals between October 1, 2020 and October 1, 2021. We included all authors whose names and emails were available and all article types/study designs. For full details on our semi-automated approach to extracting author emails, please see our search strategy in S1 File.

Participant recruitment

The survey was sent only to those researchers identified via our sampling procedure (i.e., a closed survey). Potential participants received an email containing a recruitment script which detailed the purpose of the study and invited them to review our informed consent form and complete our anonymous online survey. Participation in the survey served as implied consent; we did not require signed consent to maintain anonymity. To send emails to the sample of authors, we used mail merge feature in Microsoft 365. This tool allows for the personalization of emails without having to individually customize and send each out. In the case of non-response, we sent 3 reminder emails to potential participants at weekly intervals after the initial invitation. We closed the survey 4 weeks after the initial invitation. We did not provide any specific incentive to complete the survey.

Survey

The full survey is available in S2 File. The survey was administered using SurveyMonkey and could be completed in about 10 min. The survey contained a total of 19 questions starting with 4 demographic questions about the participants, including their gender, research role, research area, and country of residence. Participants were then asked to complete questions about their perceptions of reproducibility in biomedicine, questions about their experience with reproducibility, and questions about perceptions of barriers and facilitators to conducting reproducibility projects. The survey questions were presented sequentially and using adaptive formatting to present only certain items based on the participant’s response to previous questions. Most questions were multiple choice, with 2 questions asking participants to expand on their responses using a free-text box. The survey was purpose-built for the study by the research team, building directly off the previously published Nature reproducibility survey [10]. We included several of the previous study’s questions directly in this study, modified some slightly, and made some more specific to the biomedical research setting. We also introduced some novel questions on reproducibility. The survey was pilot tested by 3 researchers (not on the team) to ensure clarity and acceptability of format and we edited the survey to address their feedback regarding the clarity of survey questions and issues with illogical survey flow. Participants were able to skip any question.

Data management and analysis

Data were exported from SurveyMonkey and analyzed using SPSS 28 (RRID:SCR_002865). We report descriptive statistics including count and percentages for all quantitative items. For the qualitative items, we conducted a thematic content analysis. To do so, 2 researchers individually read all text-based responses and assigned a code to summarize the content of the text. This inductive coding approach involves creating codes based on the data itself, rather than from a pre-established coding framework. Codes were refined iteratively upon exposure to each text-based response read. After discussion to reach consensus on the codes used, we then grouped the agreed codes into themes for reporting in tables. Coders were members of the project team who were not blinded to the study aims.

Results

Protocol amendments

In our original protocol, we said we would take a random sample of 1,000 journals from MEDLINE and extract information from the first 20 authors. This approach required extensive manual extraction, so we opted to restrict our random sample to 400 journals and semi-automate extraction of author information for an entire year’s worth of publications. Our revised method meant that we obtained and used listed emails from all authors on an identified paper (i.e., we were not restricted to corresponding authors).

Demographics

A total of 24,614 emails were sent, but bounce backs were received from 1380, meaning 23,234 emails were sent successfully to potential participants. A total of 1,924 participants accessed our survey, of whom 1,630 participants provided completed responses (response rate 7%; this frequency is slightly lower than the estimated 1,800 responses reported in our protocol). Most participants were Faculty Members/Primary Investigators (N = 1,151, 72%) and more than half of participants were male (N = 943, 59%). Respondents were from more than 80 countries, with the USA (N = 450, 28%) having the highest representation. About half of participants reported working in clinical research (N = 819, 50%). Further demographic details by role, gender, country, and research area are provided in Table 1.

Perceptions of reproducibility

When asked whether there was a reproducibility crisis in biomedicine most researchers agreed (N = 1,168, 72%), with 27% (N = 438) indicating the crisis was significant and 45% (N = 703) indicating a slight crisis (we note that a “slight crisis” is a bit of an oxymoron, but retained the wording from the original Nature survey for comparison purposes); see Fig 1 and S3 File for breakdown by discipline. Compared to the previously published Nature study (N = 819, 52%), fewer participants in our study felt there was a “significant reproducibility crisis” (N = 438, 27%). This difference was even larger when we restricted to the Nature study participants who indicated they worked in medicine (N = 203, 60%).

thumbnail
Fig 1. Participant perceptions of a reproducibility crisis.

Data is presented overall for all participants in the current study and is broken down by research focus area in medicine. Results are presented in context to the overall Nature study findings and specifically to participants from this study indicating they worked in medicine. The underlying data for this figure can be found at https://osf.io/dbh2a.

https://doi.org/10.1371/journal.pbio.3002870.g001

Participants were then asked what percentage of papers in each of biomedical research overall, clinical biomedical research, in vivo biomedical research, and in vitro biomedical research they thought were reproducible, see Fig 2. Only 5% (N = 77) thought more than 80% of biomedical research was reproducible. See S3 File for complete results. We provide a breakdown of responses between genders, between researchers in different biomedical research areas, and by career rank in S4 File.

thumbnail
Fig 2. Participants perceptions of the proportion of papers they think are reproducible in biomedicine overall and by biomedical research area.

The underlying data for this figure can be found at https://osf.io/dbh2a.

https://doi.org/10.1371/journal.pbio.3002870.g002

Determinants of irreproducibility

When presented with various potential causes of irreproducibility, more than half of participants responded that each presented factor contributes to irreproducibility. The top characteristic participants noted as “always contributing” to irreproducibility was pressure to publish (N = 300, 19%). Factors deemed least likely to contribute to irreproducibility were fraud (N = 320, 20%) and bad luck (N = 568, 36%). See Table 2 for complete results.

thumbnail
Table 2. Participant perceptions of the causes of irreproducibility.

https://doi.org/10.1371/journal.pbio.3002870.t002

A total of 97 (6%) participants provided a written response to elaborate on what they perceived were causes of irreproducibility. Responses were coded into 16 unique codes and then thematically grouped into 7 categories: ethics, research methods, statistical issues, incentives, issues with journal and peer review, lack of resources, and other. For definitions and illustrative examples of the codes, see Table 3.

thumbnail
Table 3. Thematic analysis of perceived causes of irreproducibility.

https://doi.org/10.1371/journal.pbio.3002870.t003

Experiences with reproducibility

Participants were asked about their experience conducting reproducibility projects. Nearly a quarter of participants indicated that they had previously tried to replicate one of their own published studies and failed to do so (N = 373, 23%), whereas 31% (N = 501) indicated all such replications they had conducted had yielded the same result as the original, and slightly less than half of participants indicated that they had never tried to replicate any of their own published work (N = 734, 46%). Among the 874 participants who indicated they had tried to replicate one of their own studies, when asked to consider their most recent replication effort, 313 (36%) indicated they had published the results.

Almost half of the participants indicated that they had previously tried to replicate a published study conducted by another team and failed to do so (N = 724, 47%), whereas 10% (N = 156) indicated all replications that they had attempted were successful, while 43% (N = 666) indicated they had never tried to replicate someone else’s published research. Among those who had published their replication study, when asked to consider their most recent replication effort, 29% (N = 224) indicated it took about the same amount of time to publish as a typical non-replication paper. A quarter (N = 189, 25%) of participants who had attempted to replicate others’ research indicated they had no plans to publish their replication study. Eighty-five percent of participants (N = 1,316) indicated they had never been contacted by another researcher who was unable to reproduce a finding they previously published. See Table 4 for complete results.

A total of 724 participants responded to the item about why they replicated their own research, of which 675 (93%) provided relevant text-based responses. Responses were coded into 17 unique codes and then grouped into 7 themes: training, ensuring reproducibility, additional research, addressing concerns, joining/setting up a new lab, for publication or due to peer review, and other. For illustrative examples of the codes, see Table 5.

thumbnail
Table 5. Thematic analysis of reasons why participants replicated their own study.

https://doi.org/10.1371/journal.pbio.3002870.t005

A total of 748 participants responded to the item about why they replicated someone else’s research, of whom 700 (94%) provided relevant text-based response data. Responses were coded into 19 unique codes and then grouped into 7 themes: trustworthiness, extending and improving research, application to new setting, new research, interest, training, and other. For illustrative examples of the codes, see Table 6.

thumbnail
Table 6. Thematic analysis of reasons why participants replicated someone else’s study.

https://doi.org/10.1371/journal.pbio.3002870.t006

Support for initiatives to enhance reproducibility

Few participants reported that their research institution has established procedures to enhance reproducibility of biomedical research (N = 254, 16%), and almost half reported that their institutions did not provide training on how to enhance the reproducibility of research (N = 731, 48%) with an additional 503 (33%) reporting they were unsure of whether such training existed. We asked participants to provide information and links to relevant reproducibility training at their institution, which resulted in information or (functioning) links to 24 unique training resources (see 5). Among these 24 sources, just 9 (38%) clearly described specific openly available (e.g., no paywall) training related to reproducibility. Most researchers responded that they perceived their institution would value them doing new biomedical research studies more than replication studies (N = 1031, 67%). The majority also indicated that it would be harder to find funding to conduct a replication study than a new study (N = 1,258, 83%), with just 7% (N = 112) indicating they were aware of funders providing specific calls for conducting reproducibility related research. For full results, see Table 7.

thumbnail
Table 7. Participants perceived support for reproducibility.

https://doi.org/10.1371/journal.pbio.3002870.t007

We asked participants to indicate how much they agreed with the statement “I feel I am more concerned about reproducibility than the average researcher in my field and at my career stage” as a way to indirectly address potential bias in self-selection to complete the survey. Participants responded on a 5-point scale with endpoints, strongly disagree (1) and strongly agree (5). Participants reported a mean response of 3.2 (N = 1,402, SD = 0.89) that corresponds to the mid-point of the scale “neither agree nor disagree.”

Discussion

We report the results of an international survey examining perceptions of reproducibility. Almost 3 quarters of participants reported that they felt there was a reproducibility crisis in biomedicine. The concern appears to apply to biomedicine overall, but also specifically to clinical research, in vivo research, and in vitro research (11% or fewer participants indicated that they think more than 80% of papers in each category were reproducible). Researchers agreed that a variety of factors contribute to irreproducibility; however, the chief characteristic that most participants indicated “always contributes” to irreproducible research was a pressure to publish. Concerns about how the current system of academic rewards stresses quantity over quality have been expressed for decades [1416]—a sentiment supported by this study’s data, which suggests that researchers’ performance is negatively impacted, in terms of producing reproducible research, by what the academic system incentivizes.

More than half of participants reported having tried to replicate their own work previously, with almost a quarter indicating that when they did so they failed, and many indicating that they do not intend to publish their findings. Similar findings were reported when asked about whether participants had tried to replicate another researcher’s study, with 57% indicating they had done so, and 47% indicating the replication failed. The majority of participants had not been contacted by another researcher who was unable to reproduce their findings, which suggests that teams of researchers attempting to reproduce studies do not typically communicate despite the potential value for this contact to enhance reproducibility [9,17].

We asked several items about researchers’ perceptions of their institution’s support for reproducibility (Table 7) and the findings collectively suggest gaps in incentives and support to pursue reproducibility projects. For example, with just 16% of respondents reporting awareness that their institution has established procedures to enhance reproducibility, and 67% of researchers perceiving their institution values novelty over replication, our results suggest that overall, researchers perceive that institutions are not doing their part to effectively support and reward research reproducibility [18]. The growth of “Reproducibility Networks,” national peer-led consortiums aiming to promote reproducibility and understand factors related to irreproducibility, are a promising opportunity to rectify this situation [19]. For example, the UK Reproducibility Network (UKRN) provides an opportunity for institutions to formally commit to reproducibility by joining the network and requires a formal role within the senior management team of each member institution to support the delivery of network activities with the institution. The UKRN boasts a range of events and shared resources to support reproducibility [20]. The structure of reproducibility networks allows for harmonization but also flexibility in approach. Obtaining a sustained funding mechanism will be critical to their growth and ongoing success in terms of their long-term impact.

This study built off of an earlier study of more than 1,500 researchers surveyed by Nature about reproducibility [10]. The current study differs in several important ways. Firstly, the focus is exclusively on biomedicine, since to our knowledge no large scale and representative survey of biomedical researchers has been conducted to date. Indeed, just 203 (13%) of the 1,576 researchers who completed the original study indicated “medicine” as their main area of interest. Secondly, we randomly sampled researchers from publication lists, meaning we are able to report a response rate. This was not possible in the Nature survey, which was emailed to Nature readers and advertised on affiliated websites and social media outlets, meaning that the number of individuals encountering the survey is unknown. While it is possible that among those invited to take part there is bias among participants who choose to complete the survey, our approach has been chosen to help minimize surveying those explicitly active in reproducibility projects or related initiatives. Indeed, the finding that overall participants report not to differ from their belief of their peers regarding their level of concern about reproducibility provides some assurance that our sampling strategy was effective. We also find there is not much difference between different groups’ responses (S4 File).

We acknowledge several limitations in our approach. Firstly, our study survey was purpose-built, and while drawing on the Nature study, was not designed according to a particular theoretical framework. This type of approach lends itself well to the exploratory and descriptive goals of the current study; however, it makes it harder to test hypotheses or measure changes over time. A validated scale to assess perceptions of reproducibility and its cause would be of value to the community for future research. We did not look at correlations between responses to items on our survey, as doing so would be a post hoc change from our protocol. Future research using a validated scale may be better positioned to consider potential relationships between responses a priori (i.e., how one’s perception of a reproducibility crisis relates to the perceived proportion of research that is reproducible). Future surveys would also benefit from a more extensive approach to piloting for clarity and readability. Another potential limitation is that the coders conducting the thematic analysis were not blind to the study’s objective (i.e., were members of the research team). While the study is descriptive it remains possible that knowledge of the study aim and expertise in the reproducibility literature biased the coding approach. Moreover, only a small number of participants (6%) provided text-based descriptions to nuance their survey responses. It is perhaps not surprising given that the survey offered no incentive to take part, but it raises the possibility that those that took the time to respond to text-based items may differ from those who did not. Indeed, the top 3 most represented countries in our sample overall were the US, Canada, and the UK. While we had over 80 countries represented, our responses tended to be Western-centric which may present another bias, even in these countries tend to be among the most productive producers of research. Future research delving more specifically into recruiting from a specific country, or set of countries, could help improve our understanding of how national contexts may lead to differences in perceptions and practices related to reproducibility. Finally, our sampling approach allowed us to obtain any listed emails on a given study article. This means that more than one of the authors of a given article in our random sample may have responded to the survey. While our random sampling approach provided a strong basis for sampling, this nuance presents the possibility for some bias as authors of different studies may vary more considerably and we have no way of knowing how survey participants were represented within or between given papers. While it may be tempting to conduct comparison testing of our results compared to the earlier Nature replication survey, we are not positioned to do so. Our study did not aim to replicate the original Nature paper and was registered as a descriptive study. Given that we sampled exclusively biomedical researchers, and that our recruitment approach for sampling differed significantly, it would be very difficult to disentangle whether differences in results reflect sampling bias in the original study, temporal changes in the research ecosystem over time, different perceptions in biomedicine compared to research more broadly, or a combination of these factors. We know of no relevant framework that would have been appropriate to implement in order to test specific hypotheses within the biomedical community. Further, null hypothesis testing is not appropriate for determining beliefs and when hypothesis testing is used you need a formal sample size calculation [21,22].

Concerns about reproducibility are being widely recognized within research but also more broadly in the media [23]. The COVID-19 pandemic has shifted our thinking about research transparency [24] and highlighted issues with fraud in research, poor-quality reporting, and poor-quality study design [25,26], all factors that can contribute to irreproducible research. As stakeholders work to introduce training and interventions to enhance reproducibility, it will be critical to monitor the effectiveness of these interventions. This international survey provides a contemporary cross-section of the biomedical community. While our survey approach and direction of findings are consistent with the previous Nature study, ongoing monitoring of perceptions of reproducibility in the community is critical to gauge shifts over time. Indeed, conducting this same survey again in the future would allow for a temporal comparison on how perceptions and behaviors shift over time. This suggestion aligns with the approach of the EU Commission who in their report “Assessing the reproducibility of research results in EU Framework Programmes for Research” [27], describe an approach to understand and monitor the progress of reproducibility over time. In the work, researchers were also asked about their perceptions of reproducibility and 60% agreed that there is a reproducibility crisis, but also overwhelmingly agreed (82%) that reproducibility is “important” or “very important” in their discipline. Like our findings, participants indicated several key challenges to achieve reproducibility, with issues around cultural factors being paramount. Our finding that a pressure to publish was most likely to be rated as “always contributing” to irreproducibility is consistent with the conclusions of this report. Collectively, the outcomes of this body of evidence highlight perceived causes and constraints to producing reproducible research which should be prioritized within the biomedical community and targeted with interventions and supports to create improvements over time. An important consideration of future survey or reproducibility monitoring will be to be explicit about the language and definitions of reproducibility used—terminology in this space is complex and the same words can reflect different concepts between researchers within and between research areas. Clear definitions for reproducibility will help to ensure that differences observed in survey responses over time, or between groups, reflect more than simply differences in how surveys are interpreted.

Supporting information

S1 File. Search strategy used to obtain a random sample of journals to extract author contact information from to identify possible participants.

https://doi.org/10.1371/journal.pbio.3002870.s001

(DOCX)

S2 File. Survey administered to participants.

The survey was administered online and used survey logic to present relevant items.

https://doi.org/10.1371/journal.pbio.3002870.s002

(DOCX)

S3 File. Supplementary tables presenting a comparison of perceptions of reproducibility between our findings and the original Nature paper by Baker, presented overall and by discipline.

In addition, we present participant perceptions of reproducibility in different research areas.

https://doi.org/10.1371/journal.pbio.3002870.s003

(DOCX)

S4 File. Additional analyses of survey responses by gender and academic role.

https://doi.org/10.1371/journal.pbio.3002870.s004

(DOCX)

S5 File. List of training available on reproducibility mentioned by survey participants.

https://doi.org/10.1371/journal.pbio.3002870.s005

(DOCX)

References

  1. 1. Buck S. Solving reproducibility. Science. 2015;348:6242. pmid:26113692
  2. 2. Munafò MR, Nosek BA, Bishop DVM, et al. PERSPECTIVE A manifesto for reproducible science. Nat Hum Behav. 2017;1(January):1–9. pmid:33954258
  3. 3. Collins FS, Tabak LA. NIH plans to enhance reproducibility. Nature. 2012;505(7485):612–613. pmid:24482835
  4. 4. Begley CG, Ioannidis JPA. Reproducibility in science: Improving the standard for basic and preclinical research. Circ Res. 2015;116(1):116–126. pmid:25552691
  5. 5. Goodman SN, Fanelli D, Ioannidis JPA. What does research reproducibility mean? 2016;8(341). pmid:27252173
  6. 6. Makel MC, Plucker JA. Facts Are More Important Than Novelty: Replication in the Education Sciences. Educ Res. 2014;43:304–316.
  7. 7. Hardwicke TE, Thibault RT, Kosie JE, Wallach JD, Kidwell MC, Ioannidis JPA. Estimating the Prevalence of Transparency and Reproducibility-Related Research Practices in Psychology (2014–2017). Perspect Psychol Sci. Published online March 8, 2021:1745691620979806. pmid:33682488
  8. 8. Hardwicke TE, Wallach JD, Kidwell MC, Bendixen T, Crüwell S, Ioannidis JPA. An empirical assessment of transparency and reproducibility-related research practices in the social sciences (2014–2017). R Soc Open Sci. 2020;7(2):190806. pmid:32257301
  9. 9. Chang AC, Li P. Is Economics Research Replicable? Sixty Published Papers from Thirteen Journals Say”Usually Not”. 2015.
  10. 10. Baker M. Is there a reproducibility crisis? Nature. 2016;533:452–454.
  11. 11. Le Noury J, Nardo JM, Healy D, et al. Restoring Study 329: Eficacy and harms of paroxetine and imipramine in treatment of major depression in adolescence. BMJ. 2015:351. pmid:26376805
  12. 12. Ending Medical Reversal | Johns Hopkins University Press Books. Accessed December 7, 2021. Available from: https://jhupbooks.press.jhu.edu/title/ending-medical-reversal.
  13. 13. OSF | Reproducibility in biomedicine: A cross-sectional survey. Accessed June 27, 2023. Available from: https://osf.io/3ksvz/.
  14. 14. Fanelli D. Do pressures to publish increase scientists’ bias? An empirical support from US states data. PLoS ONE. 2010;5(4). pmid:20422014
  15. 15. Garfield E. What is the primordial reference for the phrase “publish or perish”? Forensic Sci. 1996;10(12):11.
  16. 16. Home. DORA. August 1, 2023. Accessed March 1, 2024. Available from: https://sfdora.org/.
  17. 17. Nosek BA, Errington TM. What is replication? PLoS Biol. 2020;18(3):e3000691. pmid:32218571
  18. 18. Begley CG, Buchan AM, Dirnagl U. Robust research: Institutions must do their part for reproducibility. Nature. 2015;525(7567):25–27. pmid:26333454
  19. 19. From grassroots to global: A blueprint for building a reproducibility network | PLOS Biology. Accessed May 25, 2023. Available from: https://journals.plos.org/plosbiology/article?id=10.1371/journal.pbio.3001461. pmid:34758016
  20. 20. UK Reproducibility Network. Accessed June 27, 2024. Available from: https://www.ukrn.org/.
  21. 21. Frick RW. The appropriate use of null hypothesis testing. Psychol Methods. 1996;1(4):379–390.
  22. 22. Hernán MA, Greenland S. Why Stating Hypotheses in Grant Applications Is Unnecessary. JAMA. 2024;331(4):285–286. pmid:38175628
  23. 23. Most scientists “can’t replicate studies by their peers.” BBC News. Available from: https://www.bbc.com/news/science-environment-39054778. February 22, 2017. Accessed May 25, 2023.
  24. 24. Sharing Research Data and Findings Relevant to the Novel Coronavirus (COVID-19) Outbreak. Wellcome Trust. Available from: https://wellcome.ac.uk/coronavirus-covid-19/open-data.
  25. 25. Abritis A, Marcus A, Oransky I. An “alarming” and “exceptionally high” rate of COVID-19 retractions? Account Res. 2021;28(1):58–59. pmid:32634321
  26. 26. Glasziou PP, Sanders S, Hoffmann T. Waste in covid-19 research. BMJ. 369. pmid:32398241
  27. 27. Athena RC, Directorate-General for Research and Innovation (European Commission), Know-Center, PPMI. Assessing the Reproducibility of Research Results in EU Framework Programmes for Research: Final Report. Publications Office of the European Union; 2022. Accessed June 27, 2024. Available from: https://data.europa.eu/doi/10.2777/186782.