Skip to main content
Browse Subject Areas

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Improving quality of preclinical academic research through auditing: A feasibility study

  • Claudia Kurreck,

    Roles Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing

    Affiliation Department of Experimental Neurology, Charité - Universitätsmedizin Berlin, Berlin, Germany

  • Esmeralda Castaños-Vélez,

    Roles Writing – review & editing

    Affiliation NeuroCure Cluster of Excellence, Charité - Universitätsmedizin Berlin, Berlin, Germany

  • Dorette Freyer,

    Roles Methodology

    Affiliation Department of Experimental Neurology, Charité - Universitätsmedizin Berlin, Berlin, Germany

  • Sonja Blumenau,

    Roles Methodology

    Affiliation Department of Experimental Neurology, Charité - Universitätsmedizin Berlin, Berlin, Germany

  • Ingo Przesdzing,

    Roles Methodology, Software

    Affiliations Department of Experimental Neurology, Charité - Universitätsmedizin Berlin, Berlin, Germany, QUEST Center for Transforming Biomedical Research, Berlin Institute of Health, Berlin, Germany

  • Rene Bernard,

    Roles Writing – review & editing

    Affiliations Department of Experimental Neurology, Charité - Universitätsmedizin Berlin, Berlin, Germany, NeuroCure Cluster of Excellence, Charité - Universitätsmedizin Berlin, Berlin, Germany, QUEST Center for Transforming Biomedical Research, Berlin Institute of Health, Berlin, Germany

  • Ulrich Dirnagl

    Roles Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing

    Affiliations Department of Experimental Neurology, Charité - Universitätsmedizin Berlin, Berlin, Germany, NeuroCure Cluster of Excellence, Charité - Universitätsmedizin Berlin, Berlin, Germany, QUEST Center for Transforming Biomedical Research, Berlin Institute of Health, Berlin, Germany


How much can we rely on whether what was reported in a study was actually done? Systematic and independent examination of records, documents and processes through audits are a central element of quality management systems. In the context of current concerns about the robustness and reproducibility of experimental biomedical research audits have been suggested as a remedy a number of times. However, audits are resource intense and time consuming, and due to their very nature may be perceived as inquisition. Consequently, there is very little experience or literature on auditing and assessments in the complex preclinical biomedical research environment. To gain some insight into which audit approaches might best suit biomedical research in academia, in this study we have applied a number of them in a typical academic neuroscience environment consisting of twelve research groups with about 100 researchers, students and technicians, utilizing the full gamut of state-of-the-art methodology. Several types of assessments and internal as well as external audits (including the novel format of a peer audit) were systematically explored by a team of quality management specialists. An experimental design template was developed (and is provided here) that takes into account and mitigates difficulties, risks and systematic errors that may occur during the course of a study. All audits were performed according to a pre-defined workflow developed by us. Outcomes were assessed qualitatively. We asked for feedback from participating employees in every final discussion of an audit and documented this in the audit reports. Based on these reports follow-up audits were improved. We conclude that several realistic options for auditing exist which have the potential to improve preclinical biomedical research in academia, and have listed specific recommendations regarding their benefits and provided practical resources for their implementation (e.g. study design and audit templates, audit workflow).


Almost a decade ago, several large pharmaceutical companies blew the whistle on something that had worried the industry for some while: An exceedingly high rate of failures when trying to reproduce pivotal results of studies performed in academic laboratories, most of which were published in top tier journals [1, 2]. Since then meta-research has identified substantial deficits in the planning, execution, analysis, and dissemination of results of preclinical biomedical research [3]. It appears that preclinical biomedicine is afflicted by a systemic quality problem [4]. A number of remedies have been suggested, some of which appear to gain traction [5] in increasing transparency and data sharing.

Rooted in the beginnings of modern science as an affair of “gentlemen” [6], todays research enterprise is still based largely on trust among scientists [7]. However, given the hyper- competitiveness of the academic system, how much can we rely on the fact that what was reported was actually done? Or exclude that things were done that were not reported? Any laboratory can claim that processes have been implemented to safeguard research quality, for example methods to minimize bias such as randomization and blinding, or the adherence to Standard Operating Procedures (SOP). However, such claims need to be verifiable to be useful to the community. It needs to be demonstrated that what should be done is being done.

Funders and journals increasingly request compliance with their own [8, 9] or general standards, such as making data publicly available, or complying with reporting guidelines such as ARRIVE [10]. Unfortunately, the available evidence suggests that these measures are of limited success [1114]. It appears that self-reporting of adherence to research quality standard often fails to reflect actual improvements in research practice.

Structured quality management systems (QMS) [15, 16] and more relevant to biomedicine, Good Clinical Practice (GCP) methodology [17] require auditing to avoid discrepancies between reporting and what was actually done in a study. Audits of academic non-clinical research have been suggested a number of times [1820]. A recent computer simulation suggests that random auditing of research groups can improve research quality, and is highly cost effective [6]. In addition to safeguarding Good Scientific Practice (GSP), auditing may bring additional benefits such as promoting scientific collaboration, the development of common protocols, or fostering communication, transparency and continuous improvement of the scientific processes.

Besides lack of funding, expertise, and a dearth of suitable systems, the prospect of auditing may be among the chief reasons why academia has so far shunned QM systems altogether. In our own experience with ISO 9001 [21], the audits performed by the certification agency were disappointing: Research specific deficiencies in our quality management remained undetected by the auditors untrained in preclinical research, and targeted advice how to improve research quality was consequently very limited.

The Department of Experimental Neurology with its diverse methodological research structure at one of the largest research hospitals in Europe is in an ideal position to investigate whether the controversial instrument of auditing can be practically and effectively applied to an academic research environment to hold researchers more accountable to quality standard in their daily work. Therefore, we have developed a QM system tailored to research [22], in which different forms of assessments and audits are investigated. Specific criteria were favorable cost-benefit ratio, acceptance by our scientists, but also potential efficiency and applicability for other academic biomedical research institutions.

In the following, we will first review which forms of audits we considered suitable for academic preclinical biomedical research and therefore were tested at our department over a time period of several years. Preliminary results of those tests are then presented, and limitations of our approaches discussed.


Site were audits were performed

The research environment of the Department of Experimental Neurology consists of twelve research groups with about 100 researchers, students and technicians. They study fundamental mechanisms of cerebrovascular physiology and pathophysiology using a variety of state-of-the-art technologies and approaches, including molecular biology, cell biology, biochemistry, ultrahigh field magnetic resonance imaging, histology, multi-photon microscopy, in vivo modelling of diseases in rodents, cell cultures and brain organoids, and others.

Definitions of audits and assessments

In research, an audit is a systematic and independent examination of records, documents and processes of an organization (e.g. university, department, research group) to ensure compliance to pre-specified requirements or criteria. An assessment is a demonstration that specified requirements relating to a product, process, system, person or body are fulfilled as defined in ISO 17000 [23].

Methodology for conducting audits and assessments

A number of different audits and assessment methods with various prerequisites and objectives are in principle relevant at the academic preclinical research setting [23, 24].

Audits and assessments were divided into three categories: self-assessments, internal and external audits. Since every research environment is different, we selected certain types of audits suitable (Table 1), performed different types of audits and assessments (Table 2), while we declared other types as unfit (Table 3) to our specific research settings.

Table 1. Types of audits and assessments performed at the Department of Experimental Neurology.

For a more detailed description of specific assessments and audits, see Table 2.

Table 2. Description of the assessments and audits listed in Table 1.

Table 3. Examples for types of audits and assessments, which were considered, but not performed at the Department of Experimental Neurology.

Over three years we have carried out and evaluate different forms of audits and assessments, to define, which of the audit forms are suitable for basic biomedical research. During the first three years working according to DIN EN ISO 9001:2008, we followed the required external certification and monitoring audits, the internal ring audits and the self-assessment with the FMEA [25]. After the change of the QMS from ISO 9001 to PREMIER, other forms of audits and assessments were tried out, such as error reporting using the LabCIRS [24], internal audits of methods, data, documents and processes. The peer audit was performed as a new form of external audit specially tailored to the non-regulated academic environment. The collection of errors and critical incidents with LabCIRS is an ongoing process. The internal audits have been carried out in line with current topics and the need for improvements and standardization of specific processes in our particular research environment.

Inclusion and exclusion criteria

Before conducting audits and assessments we defined inclusion and exclusion criteria:

Inclusion criteria:
  • Willingness to participate
  • Availability of relevant expertise in the scientific area to be audited
  • SOPs available and in use for at least 3 months before the audit
Exclusion Criteria:
  • Unwillingness to participate
  • Insufficient proficiency in the audited area

Declaration of ethics

The conduct of the audits did not involve individuals. It dealt with research issues, such as data storage or implementation of methods, so that all data from the audits conducted were anonymous at any time. The assessments and reviews reported in our communication are standard quality management procedures as they are carried out in millions of companies and institutions worldwide, and as such are not subject to IRB approval. As a matter of principle, the Charité IRB does not make any statements about non-competence, i.e. there is no need for IRB approval. Audit participants provided no informed written consent.

Table 1 lists all types of audits and assessments performed at the Department of Experimental Neurology as part of our ISO 9001 certification and later quest to establish a bespoke academic QM system. Self-assessments were carried out in the form of risk and error management, as well as internal audits such as method, data, documentation and process audits. A special form of internal auditing were ring audits of the Charité, which were carried out in preparation for the external audits. External audits according to ISO 9001 were conducted in the form of certification and monitoring audits. As a novel form of external audits, a peer audit was devised and performed by experts from another scientific research team (German Mouse Clinic, GMC, Munich Germany). Peer audits are more informal, less structured and aim to evaluate the actual rather than the formal performance.

Table 2 further details the methods used to carry out these audits, the audit objectives, and the frequency and the time point when they were performed.

Template for experimental planning of the audited experiment or study.

The Department of Experimental Neurology has developed an experimental design template that takes into account and mitigates difficulties, risks and systematic errors that may occur during the course of a study. Specific questions in the template are geared to reduce the risk of bias in the experiments and raise awareness for project-specific quality issues. This template was developed, used and tested during the peer audit (see Table 2).

Template audit plan.

An audit plan must be prepared before each audit. In advance, the subjects to be audited and the corresponding schedule are agreed with all participants (see S1 File). In terms of transparency and reliability, it is important to comply with the audit plan, which is the responsibility of the auditing persons.

Types of audits and assessments, which are not performed at the Department of Experimental Neurology.

Quality management systems use several types of audits and assessments, often to maintain certification and accreditation. Each organization defines which types are best suited to its needs and establishes corresponding audit program objectives. Most available systems are set out to optimize products, services, customer- and supplier relationships all of which have limited applicability in the preclinical academic context. Therefore, several types of audits, such as customer or supplier audits, were not feasible in our environment because we work in a non-regulated, academic, pre-clinical context. We focused only on those considered applicable to our environment in order to identify and analyze potential gaps (such as in protocols, in the implementation of methods, in the storage of raw data etc.).

Table 3 provides a selection of audit types that might be of interest in biotechnological laboratories for example, but are not useful in the biomedical research we conduct.

Audit workflow.

All types of audit specified in Table 2 were performed according to a pre-defined workflow developed by us (see Fig 1). Audits always included four main steps: Preparation, Execution, Evaluation & Reporting and Proof of Effectiveness. This quality assurance workflow allows the close support of a specific project as well as an entire organization, thus making the improvement at different levels in preclinical research possible. It should be noted that the performance of i.e. internal audits or proof of effectiveness is not necessarily required by professional auditors or QM personnel, but can also be carried out by scientists who are familiar with the audit topic.

Fig 1. Audit workflow.

This audit workflow describes the four steps performed during an audit in an academic preclinical research environment, but is universally applicable to all areas in which audits are performed. This audit workflow is a practical example of Deming’s PDCA (Plan-Do-Check-Act) cycle, which incorporates the process of continuous improvement.


As integral part of structured quality management in our department, we carried out and evaluated different types of audits and assessments (such as methods-, data-, process- or peer audits) in an environment typical for preclinical academic research (for details see Tables 1 and 2 above).

Audits were considered successful if they were accepted by the employees, conducted in a clear and non-aversive language, and if the scope and objectives of the audit were clearly defined and if the measures resulting from the audits proved to be effective. Acceptance we defined as agreement of the employees with the objective, focus and manner in which the audit was carried out. The audits were not performed in the style of an examination, but presented as an opportunity for improvement and further development of problems and difficult issues.

The acceptance of the audits was assessed qualitatively. We asked for feedback from participating employees in every final discussion of an audit and documented this in the audit reports. Based on these reports we improved the follow-up audits.

Through our practical experience, we are able to make recommendations for types of audits suitable for preclinical research, which are accepted by researchers, technicians and students. Suitability refers to their importance and relevance in the non-regulated research environment, compliance with regulatory requirement, and the generation of robust results. A summary of the advantages and disadvantages of each type of audit can be found in Table 4. A graphical representation of the number of advantages and disadvantages of audits and assessments see Fig 2.

Fig 2. Graphical representation of the number of advantages and disadvantages of audits and assessments as perceived by members of the department.

Table 4. Results of audits and assessments performed in the Department of Experimental Neurology.


With respect to self-assessment, failure and error management with a structured anonymous error and critical incidence reporting system (LabCIRS [24], freely available at was effective in improving research quality. This was reflected in the effectiveness of the measures resulting from the error reports in LabCIRS, the low rate of recurrence of the same error and increased awareness of specific topics. These included, for example, improvements in sample identification, comprehensive documentation of experiments in an electronic laboratory book, and the calibration of pipettes. Within three years, a transparent error culture evolved which was built on the overwhelmingly positive experience and acceptance of discussing errors and their prevention. Interest in our LabCIRS has been growing over time, and meanwhile has not been only adopted by other laboratories but is now centrally offered to anyone in the Charité and MDC (LabCIRS BIH) research environment.

We found that the various forms of audits differed greatly with respect to requirement of resources, user experience, efficiency, and fitness for purpose. Interestingly, some of the audit types, such as the Failure Mode and Effects Analysis (FMEA), despite their popularity in other professions and settings (e.g. in the industry or service sectors), revealed severe drawbacks in an academic research setting. For example, many risks in academic research are connected to limitations in continuity in employment contracts and research funds. Finding solutions to this institutional problem via FMEA is very limited. In our experience, FMEA proved to be time consuming, complex, and unpopular among scientists, as the many diverse discussions in the meetings showed.

Internal audits

With the help of specific and targeted internal audits, risks, errors and gaps between requirements (own, legal, standard) and actual performance were identified, analyzed, evaluated, corrected, monitored and documented. Internal assessments and audits played a key role in the implementation of structured quality assurance measures. Due to their systematic approach to specific issues, they proved to be a powerful mechanism for assessing the validity of methods and data and identifying processes and areas for refinement. With their help, it was possible to initiate a process of continuous improvement and to ensure the effectiveness of implemented measures. An example of such a process was an internal audit for the storage of primary data. Here compliance with the valid Standard Operation Procedure (SOP) and Good Scientific Practice (GSP) was checked and measures to improve the process were implemented. For storage of raw data, the responsible staff (a designated individual in each of 12 working groups) was asked, how, where and in what way the groups store their raw data. When gaps or potential problems were identified, possible solutions were discussed and measures for improvement were formulated. After approximately three months, the implementation and effectiveness of the suggested measures were evaluated. The performance of the audit is provided as an example in the audit checklist (see S2 File). We therefore recommend internal audits as a useful tool for evaluating critical processes and areas in the research environment.

Internal ring audits of the Charité were useful in preparing for the certification according to ISO 9001.

External audits

Our department underwent external auditing in two different modes, as part of the ISO 9001 certification and as independent project audit performed by an external peer group. The certification audits by an accredited agency did not have the necessary depth to investigate the operational level of our research processes. As a newly certified research unit, the main focus of the audit was on system implementation and compliance with pre-specified standards. In the follow up discussions, the scientific members of our department perceived these audits as not helpful for their research processes.

A peer audit was performed with the help of experts from the German Mouse Clinic (GMC) in Munich, who conducted an audit of a specific project at our department. Primary objective was to obtain proof of concept of the feasibility of this audit format i.e., a project peer review in a research environment. The audit included external assessment of the project design, plausibility checks, comparison of methods, scientific exchange, and evaluation of the audit format.

Preparing the audit was time-consuming. This format was new for both, the auditors and the audited, and a process had to be developed ‘from scratch’. The audit was conducted in a collegial, trusting and open atmosphere. While many audit forms extend their investigations to an entire laboratory or department, peer audits are focused on areas and scientific personnel with direct project involvement.

As opposed to other types of audit during peer audit specifics of the projects were evaluated and discussed in great depth. Using a tailored checklist valuable general and project-specific hints and tips were given. Furthermore, both sides, auditors and participants, exchanged ideas and learned from each other. It must be emphasized, however, that peer audit requires a strong commitment to transparency and academic exchange.


Audits are a central element of quality management systems. Due to the fact that structured quality management is almost unknown in academic preclinical research, there is very little experience or literature on auditing and assessments in this complex research environment. Audits can be carried out by a number of different methods, depending on objectives, scope, criteria, duration, available resources, and context. To gain some insight into which audit approaches might best suit biomedical research in academia, we have applied a number of them, and here we report our experience.

Our experience in an environment typical of biomedical research, gained in a structured way through the introduction of two different QM systems (ISO 9001 [16], and PREMIER [22]) shows that audits can be very useful in pre-clinical research, especially in the evaluation of data, methods, processes and projects. In addition, we found it advantageous for "like-minded people", i.e. colleagues or other researchers who can be trusted in terms of expertise and appropriate and reasonable performance, to conduct audits.

Depending on the size of an individual work group, laboratory, or institution, and the type of research, there are various possibilities for self-assessment, risk assessment and internal and external audits in biomedical basic research. In any case, focus and scope of the audit needs to be predefined in a structured process. To assist the implementation of internal and external audits, we here provide tools that have been useful in our environment (S1 and S2 Files, Fig 1). In our experience, audits were particularly useful in identifying potential errors and critical issues while projects and studies are planned, conducted, and analyzed. This has been demonstrated, for example, by follow-up audits, which examined the effectiveness of the measures introduced in the previous audit. In this way, a process of continuous improvement can be set in motion and quality in preclinical research improved (see Fig 1).

Peer audits as a special case for external audits are aimed at specific research projects, their methods and approaches. Due to their specificity and direct research focus, peer audits have a higher granularity than certification audits, and the findings and recommendations can be applied directly to specific research processes. We have experienced peer auditing as a bi-directional exchange on an equal footing among like-minded researchers, which ensured an open discussion culture and efficient outcome.

Finding suitable peer auditors may be difficult, since it requires expert knowledge in a specific research area or technique, an open attitude towards quality assessment, and a transparent research culture and exchange. Choice of the peers may bias the outcome of the audit. More importantly, it may be argued that fear of being scooped or loosing intellectual property may prevent scientists from opening up to auditors who are peers and thus potential competitors. First, these fears are unfounded as audits can entirely focus on non-proprietary information, and are covered by non-disclosure agreements. Audits usually revolve around methods and work which was already published by the audited laboratory or group. More importantly, however, is that what appears as a threat is actually an opportunity and a major strength of peer audits. The audited scientists as well as the auditors are united in their quest to improve their own research. During the audit, in particular if performed by a researcher working in the same field with similar methodology, both parties exchange best practice and expose pitfalls and shortcomings. This works in both directions, so that not only the audited laboratory but also the auditor benefit for their own research. In fact, this means that auditors have a selfish interest to engage in something for which a motive, besides research idealism, seems to be missing. In a sense, internal audits can be considered an ideal form of scientific discourse, more powerful than discussions at scientific conventions. Of course, for this to work both parties need to already subscribe to a certain level of openness and transparency. It is our experience that such a spirit is quite common among researchers, and that remaining skepticism as to the benefits and fear of potential threats of audits vanish once the first audit is completed. Our department has benefited from the peer audit conducted. We believe that peer audits have great potential as external reviews, as they can have an immediate positive impact on the outcome of a project or process and can build a trusting and lasting relationship between peer auditing groups.

Based on numerous informal discussions with researchers, students and technicians, regular quality meetings, and exchanges with various stakeholders outside our department, we believe that academic researchers accept audits if they are conducted in a clear and non-averse language and if the scope and objectives of the audit are clearly defined.

While our structured approach to audits has been part of our attempts to implement QM in the academic research environment, it is important to note that audits are not dependent on the existence of a QM system. Audits are a general instrument of quality assurance. Robustness and reproducibility of research results can only be achieved if reporting reflects actual practice accurately. In this context, audits are a simple and valid instrument to check the consistency between scientific practice and reporting. Quality assurance in preclinical research is based on measurement and continuous improvement to evaluate the effectiveness of the applied quality measures. To maintain a high level of performance in research work, it is necessary to react to changes in internal and external requirements and to create new solutions. Audits are one possibility to achieve these objectives.

Given the lack of structured QM in pre-clinical research, we suggest that audits should be adapted for use in this area without necessarily relying on an existing QM system. The systematic approach to QM in academic preclinical research that we are currently developing [22] includes an audit module that can be used independently and provides practical advice on how to conduct audits and assessments.


Our study has a number of weaknesses and limitations. Although we believe that our research environment is typical for a large fraction of biomedical research laboratories in academia, we can only present anecdotal evidence from one department, one area of research (neuroscience), and one country (Germany). This precludes formal statistical analysis, and unfortunately does not provide insights from other types of laboratories, such as those in the physical or chemical sciences. Our approach was also not preplanned and preregistered, and our outcome assessment is highly subjective. In addition, the number of audits performed was limited. Audits were prepared and conducted by a team of QM professionals in a research environment preconditioned to issues of QM. We nevertheless believe that our experience can be helpful to many stakeholders (researchers, institutions, funders) as to our knowledge it is the first attempt to formally approach this issue. Our results can inform larger and more systematic studies in the future and stimulate discussion.


Despite these limitations, from our hands-on experience with various approaches to auditing in preclinical academic research, and the feedback received from multiple stakeholders, we conclude the following:

  • Audits, as well as any kind of assessment can only function properly if those audited understand the concept and its merits and participate voluntarily.
  • Structured error management with a laboratory critical incidence system is an effective self-assessment activity, which can be implemented relatively easily. Beyond helping in the recurrence of reported errors, it is a powerful tool to introduce a non-punitive error culture.
  • Internal process-, document- and data audits are a straightforward approach to identify, analyze, and correct risks, errors and gaps in projects. They safeguard the validity of methods and data and help to identify areas that need improvement. These internal audits foster compliance with the complex legal and regulatory frameworks of biomedical research (animal protection laws, 3R, safety regulations, reporting guidelines like ARRIVE, etc.).
  • Depending on the size of the organization, internal audits can be done several times a year and in particular in critical stages of a project. Internal audits are also useful at the organizational level for reviewing quality assurance processes.
  • Peer audits are a promising novel tool to solicit external feedback and foster professional exchange of ongoing projects at eye-level. They are very effective in fostering the improvement of methods or processes.
  • We recommend conducting at least one peer audit during a project’s lifetime, especially if specific methodological challenges need to be solved and an exchange with colleagues is desired.
  • Auditing contributes to transparency. In general, and peer audits in particular, auditing might serve as fundamental processes of open and transparent science in the future.
  • Careful and collegial communication of expectations and goals is a prerequisite for successful audits.

Audits are an opportunity to safeguard research quality and transparency, not instruments of control. Audits can increase trust between researchers, and between researchers and the public, in particular at times where this trust has eroded. Specific forms of auditing may be selected or further developed to benefit all stakeholders in the system, researchers, institutions, funders, scholarly societies, journals, and the public. We conclude that several realistic options for auditing exist which have the potential to improve research, and have listed specific recommendations regarding their benefits and implementation. Further systematic studies are needed to select and adapt auditing approaches acceptable to and practical for all stakeholders in the academic research enterprise and evaluate how their implementation affects research quality.

Supporting information

S2 File. Checklist internal and external audit.


S3 File. Table: Baseline characteristics of participants.



We thank Valerie Gailus-Durner, Claudia Stöger, Adrian Sanz Moreno, Julia Calzada-Wack (German Mouse Clinic, GMC, Munich, Germany) for performing the peer audit and stimulating discussions.

Nikolas Offenhauser (VIOMEDO, Berlin, Germany) helped us with the ISO certification process.

We gratefully acknowledge the support of the Volkswagen Stiftung (Hannover, Germany) for support in developing the PREMIER Quality System.


  1. 1. Begley CG, Ioannidis JP. Reproducibility in science: improving the standard for basic and preclinical research. Circ Res. 2015;116(1):116–26. Epub 2015/01/02. pmid:25552691.
  2. 2. Baker M. 1,500 scientists lift the lid on reproducibility. Nature. 2016;533(7604):452–4. Epub 2016/05/27. pmid:27225100.
  3. 3. Macleod MR, Michie S, Roberts I, Dirnagl U, Chalmers I, Ioannidis JP, et al. Biomedical research: increasing value, reducing waste. Lancet. 2014;383(9912):101–4. Epub 2014/01/15. pmid:24411643.
  4. 4. Begley CG, Buchan AM, Dirnagl U. Robust research: Institutions must do their part for reproducibility. Nature. 2015;525(7567):25–7. Epub 2015/09/04. pmid:26333454.
  5. 5. Voelkl B, Wurbel H. Reproducibility Crisis: Are We Ignoring Reaction Norms? Trends Pharmacol Sci. 2016;37(7):509–10. Epub 2016/05/24. pmid:27211784.
  6. 6. Barnett AG, Zardo P, Graves N. Randomly auditing research labs could be an affordable way to improve research quality: A simulation study. PLoS One. 2018;13(4): e0195613. Epub 2018/04/13. pmid:29649314.
  7. 7. Institute of Medicine. Society’s Choices: Social and Ethical Decision Making in Biomedicine. Bulger RE, B E, Fineberg HV, editor. Washington, DC: The National Academies Press; 1995.
  8. 8. Yarborough M. Openness in science is key to keeping public trust. Nature. 2014;515(7527):313. Epub 2014/11/21. pmid:25409791.
  9. 9. Marcus E, whole Cell t. A STAR Is Born. Cell. 2016;166(5):1059–60. Epub 2016/08/28. pmid:27565332.
  10. 10. Bishop D. Rein in the four horsemen of irreproducibility. Nature. 2019;568(7753):435. Epub 2019/04/26. pmid:31019328.
  11. 11. Kilkenny C, Browne WJ, Cuthill IC, Emerson M, Altman DG. Improving bioscience research reporting: the ARRIVE guidelines for reporting animal research. PLoS Biol. 2010;8(6):e1000412. Epub 2010/07/09. pmid:20613859.
  12. 12. Leung V, Rousseau-Blass F, Beauchamp G, Pang DSJ. ARRIVE has not ARRIVEd: Support for the ARRIVE (Animal Research: Reporting of in vivo Experiments) guidelines does not improve the reporting quality of papers in animal welfare, analgesia or anesthesia. PLoS One. 2018;13(5):e0197882. Epub 2018/05/26. pmid:29795636.
  13. 13. Federer LM, Belter CW, Joubert DJ, Livinski A, Lu YL, Snyders LN, et al. Data sharing in PLOS ONE: An analysis of Data Availability Statements. PLoS One. 2018;13(5):e0194768. Epub 2018/05/03. pmid:29719004.
  14. 14. Hair K, Macleod MR, Sena ES, Collaboration II. A randomised controlled trial of an Intervention to Improve Compliance with the ARRIVE guidelines (IICARus). Res Integr Peer Rev. 2019;4:12. Epub 2019/06/18. pmid:31205756.
  15. 15. NPQIP. Did a change in Nature journals’editorial policy for life sciences research improve reporting?. BMJ Open Science. 2019;3.
  16. 16. ISO. ISO 9001:2008 Quality management systems—Requirements. ISO 9001:2008: ISO; 2008.
  17. 17. Harmonised Tripartite Guideline Good Manufacturing Practice Guide For Active Pharmaceutical Ingredients Q7 (2000).
  18. 18. Winchell T. The clinical investigator site audit—process-driven good clinical practice. Qual Assur J. 2007;11:138–42.
  19. 19. Begley CG, Ellis LM. Drug development: Raise standards for preclinical cancer research. Nature. 2012;483(7391):531–3. Epub 2012/03/31. pmid:22460880.
  20. 20. Yarborough M, Nadon R, Karlin DG. Four erroneous beliefs thwarting more trustworthy research. Elife. 2019;8. Epub 2019/07/30. pmid:31355746.
  21. 21. Lefevre IA, Balice-Gordon RJ. Learning from Principles of Evidence-Based Medicine to Optimize Nonclinical Research Practices. Handb Exp Pharmacol. 2020;257:35–54. Epub 2019/09/19. pmid:31529201.
  22. 22. Dirnagl U, Kurreck C, Castanos-Velez E, Bernard R. Quality management for academic laboratories: burden or boon? Professional quality management could be very beneficial for academic research but needs to overcome specific caveats. EMBO Rep. 2018;19(11). Epub 2018/10/21. pmid:30341068.
  23. 23. ISO. ISO/IEC 17000:2004 Conformity assessment—Vocabulary and general principles. ISO 17000:2004: ISO; 2004.
  24. 24. Dirnagl U, Przesdzing I, Kurreck C, Major S. A Laboratory Critical Incident and Error Reporting System for Experimental Biomedicine. PLoS Biol. 2016;14(12):e2000705. Epub 2016/12/03. pmid:27906976.
  25. 25. ICH guideline Q9 on quality risk management, (2009).