Skip to main content
Browse Subject Areas

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Registered Report Protocol: Survey on attitudes and experiences regarding preregistration in psychological research



Preregistration, the open science practice of specifying and registering details of a planned study prior to knowing the data, increases the transparency and reproducibility of research. Large-scale replication attempts for psychological results yielded shockingly low success rates and contributed to an increasing demand for open science practices among psychologists. However, preregistering one’s studies is still not the norm in the field. Here, we propose a study to explore possible reasons for this discrepancy.


In a mixed-methods approach, an online survey will be conducted, assessing attitudes, motivations, and perceived obstacles with respect to preregistration. Participants will be psychological researchers that will be recruited by scanning research articles on Web of Science, PubMed, PSYNDEX, and PsycInfo, and preregistrations on OSF Registries (targeted sample size: N = 296). Based on the theory of planned behavior, we predict that positive attitudes (moderated by the perceived importance of preregistration) as well as a favorable subjective norm and higher perceived behavioral control positively influence researchers’ intention to preregister (hypothesis 1). Furthermore, we expect an influence of research experience on attitudes and perceived motivations and obstacles regarding preregistration (hypothesis 2). We will analyze these hypotheses with multiple regression models, and will include preregistration experience as control variable.


Ever since Ioannidis [1] argued that most published research findings are false and a multitude of attempts failed to replicate previously significant effects (e.g., [2]), the reliability of published research findings has been a subject of discussion across many scientific disciplines. Summarizing these concerns, the term replication crisis [3] arose, where replicability refers to the attempt to repeat an experiment to re-test the original effect [4]. When Nature [5] conducted a survey of more than 1500 researchers of multiple disciplines, 70% of researchers reported that they had failed to replicate studies by others, and more than 50% had failed to replicate their own studies. Overall, 90% of researchers indicated their belief in a slight or even significant crisis [5]. In psychology, multiple large-scale research projects attempted to replicate significant effects published in top tier journals. Strikingly many attempts failed as shown by replication rates between 36 [2] and 77% [6]. Among successfully replicated effects, the effect sizes were considerably smaller than originally reported.

It has been reasoned that false positives, i.e., effects that are significant in studies but do not exist in reality, contribute to low replicability [7]. The high rate of false positive research results has largely been attributed to “questionable research practices” (e.g., see [8,9]), a collective term for any “exploitation of the gray area of acceptable practice … (which can) increase the likelihood of finding evidence in support of a hypothesis” ([9] p. 524). Examples for these practices are the failure to control for biases, selective reporting of significant results, p-hacking, or revising the hypotheses to match the results, also known as HARKing (see [8,1014 for details).

Preregistration–on the rise?

The preregistration of studies has been proposed to counter these questionable research practices (e.g., see [15], and see [10] for an overview of other open science techniques). A preregistration is a research plan that is time-stamped, created before the data has been collected or examined, and most often submitted to a public registry, thus making planned study details available to others (possibly after an embargo period) [15,16]. If the research plan changes afterwards, either a new version needs to be added or the deviations will be apparent when comparing the preregistration to the final manuscript. Thus, preregistration aims for a transparent presentation of what was planned at a certain time point and what changes may have been made to a study until its publication. Evidence from other scientific disciplines indicates that preregistration indeed works, i.e., it increases the transparency of the research process, and reduces questionable research practices and the rate of false positive findings (e.g., see [17,18]).

However, while preregistration is already well-established in other scientific disciplines and is mandated, for example, in medicine [19], it has been frequently demanded as a means to counter questionable research practices but is still not widely practiced in psychology.

On the one hand, in response to the replication crisis, many psychologists have committed themselves to the advancement and promotion of open science techniques such as preregistration (e.g., see [10,13,15,2022]). For example, Nosek et al. [20] describe preregistration as “hard, and worthwhile”, while Wagenmakers and Dutilh [22] posit “seven selfish reasons for preregistration”. Indeed, the number of preregistrations in psychology is increasing. For instance, the number of preregistrations on the Open Science Framework (OSF), a platform for sharing research materials, has been approximately doubling every year between 2012 and 2017 [21], and in a survey which was conducted in 2018, 44% of the sampled psychological researchers indicated having preregistered a hypothesis or analysis until 2017 [23].

Yet, looking at the fraction of published studies that were actually preregistered paints a different picture. In their recent study, Hardwicke et al. [24] found that only 3% of 188 examined articles from 2014 to 2017 included a preregistration statement, which contradicts the more positive outlook by [21] and [23]. Stürmer et al. [25] also found that when early career researchers were asked about questionable research practices and open science, they deemed many open science practices necessary, yet toward preregistration they expressed more reluctance: Only about half of the participants found that preregistration was fairly necessary or very necessary, and even less indicated that they planned to consider preregistering their studies in the near future.

A number of reservations are mentioned frequently when discussing preregistration in psychology, including the fear that it leaves no flexibility during study administration and eliminates the possibility to conduct exploratory analyses (as presented by e.g., [15,16]). Some people are concerned that this might stifle discovery [26]. Others worry that someone might take their preregistered and thus, publicly available, study idea and publish it before them (so-called scooping, see [27]). Additionally, the time costs and effort are often seen as obstacles regarding preregistration (e.g., see [16]). Besides these worries, some authors also express an overall critique regarding the concept of preregistration. Szollosi et al. [2830] argue that preregistration is redundant when good theories are tested, and does not itself improve theories. Others argue that preregistration cannot prevent some questionable research practices [31], or might not fit well with all types of research (see [16]). Moreover, some studies found problems with the current implementation of preregistration such as poor disclosure of deviations from preregistered plans in finished manuscripts [3234]. Although most of the listed arguments against preregistration are counter-argued by supporters of preregistration (e.g., that exploratory analyses are still possible [15,16,27]) and findings from other scientific disciplines underline its benefits (e.g., see [17,18]), these reservations persist and some researchers remain skeptical.

Aim of this survey

While previous surveys inquired about preregistration in the more general context of open science [23,25,35,36], to our knowledge no comprehensive study focusing on preregistration has been conducted. We aim to close this gap by exploring thoughts, motivations, and perceived obstacles of psychological researchers toward preregistration and how these are influenced by the time someone has worked in research or actual experience with preregistration, that is, whether someone preregistered a study in the past. For instance, we want to explore the data to find out whether the low rate of preregistrations is caused by fear of the unknown or based on negative experiences (and which ones), as well as whether the increase of preregistrations is driven by a few active supporters while others reject or are indifferent toward it. Thus, we aim to shed light on the outlined discrepancy of public support for preregistration on the one hand, and a low fraction of preregistrations on the other, while also identifying possible roadblocks for preregistration in psychology. Mixed-methods will be used, including both qualitative and quantitative approaches.

Additionally, we want to investigate two specific research questions: First, we want to examine which factors facilitate or prevent preregistration (research question 1). The theory of planned behavior [37,38] is a prolific, influential (e.g., see [39]) and well-researched (e.g., see [4045]) psychological theory that aims to predict social behavior and has been applied across various contexts (e.g., health). According to this theory, the intention to perform a behavior can be seen as a direct antecedent of the actual behavior. In this framework, the intention to preregister predicts preregistration, and how the intention is formed may be informative for effectively promoting this behavior. To our knowledge, this has not yet been studied in the context of preregistration or open science. As described by Ajzen and colleagues [37,38], three aspects influence intentions which are defined as follows: 1) Attitudes toward the behavior which result from the ratio of perceived advantages to disadvantages of performing the behavior, 2) the subjective norm which represents the perceived social pressure to perform or not perform the behavior, and 3) the perceived behavioral control which focuses on the question if the subject has the resources and skills to perform the behavior or not (also see [4045] for meta-analytical support of this model, and [46] for an overview). We will measure attitudes toward preregistration as well as subjective norms and perceived behavioral control through items in an online questionnaire, and investigate how they influence researchers’ intention to preregister their studies in the future. Based on the model’s postulations, we expect that more favorable attitudes and subjective norms as well as higher perceived behavioral control positively influence the intention to use preregistration. As the relative importance of attitudes, subjective norms and perceived behavioral control differs in dependence of considered behaviors, situations and populations [37,38], we want to test which of these is the strongest predictor for the intention to use preregistration. We will also include the perceived importance of preregistration as moderator for the strength of the influence of attitudes on intention, and the preregistration experience as a control variable. Such an extension of the model will compensate for potential non-attitudes (e.g., see [47]) and a potential sampling bias toward researchers that have already preregistered, and is explicitly allowed by the theory of planned behavior [37,46].

Regarding the intention formation, we have the following hypotheses:

  1. 1. The theory of planned behavior [37,38] can be applied to the context of preregistration to significantly predict researchers’ intention to preregister their studies in the near future, using a moderated multiple regression model. We predict that:
    1. 1.1. More beneficial attitudes are a positive predictor for the intention to preregister.
    2. 1.2. The perceived importance of preregistration moderates the effect of attitudes on intention positively.
    3. 1.3. The perceived importance of preregistration is a positive predictor for the intention to preregister.
    4. 1.4. Beneficial subjective norms are a positive predictor for the intention to preregister.
    5. 1.5. Higher perceived behavioral control is also expected to be a positive predictor.
    6. 1.6. These predictors combined can significantly predict researchers’ intention to preregister.

Second, we want to examine whether research experience predicts attitudes and the perceived intensity of motivations and obstacles (research question 2). Research experience will be operationalized as the number of years someone indicates they have worked in psychological research. Early career researchers are oftentimes seen as the driving force of the open science movement (e.g., see [48]). We want to investigate if the research experience indeed has an influence on researchers’ responses about preregistration (a similar effect was reported by Abele-Brehm et al. in a comparison of academic groups regarding hopes and fears toward open science [35]).

Regarding this second research question, the following hypotheses will be tested:

  1. 2. We predict that research experience, that is, the amount of time someone has already worked in psychological research, has an influence on attitudes, motivation, and perceived obstacles regarding preregistration. Specifically, we will conduct three multiple regressions (including preregistration experience as control variable), and we posit the following non-directional hypotheses:
    1. 2.1. Research experience is a predictor for attitudes regarding preregistration.
    2. 2.2. Research experience is a predictor for the strength of motivation to preregister.
    3. 2.3. Research experience is a predictor for how strongly obstacles to preregister are perceived.

The present survey is aimed to sample the general population of psychological researchers by recruiting participants whose articles appear on Web of Science, PubMed, PSYNDEX, and PsycInfo, as well as the subgroup of researchers who have preregistered before and who will be identified through their preregistration on OSF Registries.


Sampling procedure

Power analyses.

Data from psychological researchers at different career stages will be collected. The optimal sample size reported below has been determined by using G*Power [49,50] in combination with a thorough review of the existing literature as described in the following paragraphs. Each of the power analyses described below was specified to achieve a statistical power of 95% at a given significance threshold of 5% (α = β = .05). All power analyses are displayed in Fig 1 and are also included in the supporting information (see S1 Text).

Fig 1. Overview of the power analyses.

The detailed statistical models and corresponding power analyses are displayed for A) the moderated multiple regression that will be computed for testing hypothesis 1 and B) the multiple regressions that will be used to test hypothesis 2.

To test which factors influence the intention to preregister (scale of three items, see hypothesis 1), a moderated multiple regression model will be computed based on the rationale of the theory of planned behavior, which includes six predictors: Attitudes (scale of 24 items), perceived importance of preregistration (one item in our questionnaire), attitudes x importance, subjective norm (scale of eight items), perceived behavioral control (scale of five items) [37,38,46], and preregistration experience (one item). The theory of planned behavior has been examined using meta-analytical approaches in various contexts (e.g., health behavior). The percentage of variance of intention that was explained by attitudes, subjective norm and perceived behavioral control combined, ranged between 30.4% < R2 < 44.3% [40,4245]. We chose the lowest reported effect size (R2 = 30.4%) as minimal effect size of interest. The power analysis for the overall regression model yielded an optimal sample size of N = 55 to be able to detect the determined effect size given the effect exists (see hypothesis 1.6). Additional power analyses were conducted to compute the optimal sample size to test the individual predictors. As comparable effect sizes, R2 based on the averaged correlations of individual variables were searched for in the aforementioned meta-analyses, and the smallest ones were chosen for each power analysis. This resulted in an optimal sample size of N = 42 for testing attitudes (see hypothesis 1.1), N = 85 for testing subjective norms (see hypothesis 1.4), and N = 236 for testing perceived behavioral control (see hypothesis 1.5) as predictors for intention (with α = β = .05). Since perceived importance and its interaction with attitudes, as well as preregistration experience are not originally included in the model but added by us, no comparable effect sizes are available and thus, no power analyses have been conducted for these variables.

To investigate whether research experience has an effect on attitudes (see hypothesis 2.1), and the perceived intensity of motivations (see hypothesis 2.2) and obstacles (see hypothesis 2.3) regarding preregistration, three regression models will be computed, again including preregistration experience as control variable. As comparable effect size, the smallest reported effect size for group differences by Abele-Brehm et al. [35] was used, and the corresponding F value was used to calculate R2 for the power analyses. For these regression models, the alpha level will be corrected using the Bonferroni-Holm method to account for multiple testing. Thus, the power analyses were calculated for the smallest alpha of 1.67% (5% divided by 3). These power analyses suggested an optimal sample size of N = 296 to test the influence of research experience on attitudes (see hypothesis 2.1), perceived intensity of motivations (see hypothesis 2.2), and perceived intensity of obstacles (see hypothesis 2.3) regarding preregistration.

As the optimal sample size for the regression models to test hypothesis 2 (N = 296) is the highest computed necessary N, it will constitute the targeted sample size.

Data collection and stopping rule.

To ensure that our sample contains researchers with varying research experience, a quota sampling will be used to evenly collect data from researchers with different degrees, that is, bachelor’s degree, master’s degree, doctoral degree, or habilitation and/or full professorship. Specifically, a quota sampling will be used with a 25% quota for each subgroup. Yet, as it might be the case that some quotas might not be filled, quotas are defined for a slightly bigger sample than the necessary N = 296 to compensate for this eventuality. In particular, N = 400 will be the basis for our quotas which will therefore be n = 100 for each group. This enables us to reach the a priori computed power even if not all quotas can be filled, while still ensuring that no group is overrepresented.

To recruit this sample, a method similar to the one by Field et al. [51] will be used. The term “psychology” will be searched for on specified databases, and resulting hits will be sorted from newest to oldest. These documents will be scanned for authors whose email addresses can be found via institutional or personal websites linked to the respective work, or by searching for the author via Google, Google Scholar, and ResearchGate. Duplicated email addresses that may be sampled from different databases will be excluded. Identified persons will be invited to participate in the survey via email.

For this search, the databases Web of Science (, PubMed (, PSYNDEX (, and PsycInfo ( will be used to recruit a representative sample of psychological researchers.

Based on the results by Hardwicke et al. [24], we anticipate that only a small proportion of the research articles that will be found on the general databases will include a preregistration statement, which might lead to a small number of participants with any preregistration experience. Thus, we decided to send 10% of invitations to authors of preregistrations in order to ensure that our survey will also include a sufficient number of participants who have preregistered before. These participants will be identified via the preregistration platform OSF Registries ( As this may introduce a sampling bias towards researchers who lean more positive towards preregistration, preregistration experience will be controlled for in the statistical analyses aimed to draw inferences about the general population of psychological researchers.

A customized link will be distributed among participants of each database so that the recruitment source can be inferred. The database participants were recruited from will be considered in the analyses as described in the section descriptive reports.

Response rates were rather low in many studies that relied on researchers as sample (e.g., [9,25,35,36,5153]). Based on these studies and based on the insights from our pilot study (see section pilot study), we anticipate a response rate of around 10%. To compensate for this, we will invite N = 2960 persons (n = 666 via Web of Science (22.5%), n = 666 via PubMed (22.5%), n = 666 via PSYNDEX (22.5%), n = 666 via PsycInfo (22.5%), and n = 296 via OSF Registries (10%)) to reach our target sample size of N = 296 (10% of 2960). If this does not result in a sufficient sample size, an additional invitation wave will be conducted. In this second wave, participants will subsequently be invited for quotas that have not yet been filled. For each open quota, the response rate of the first invitation wave will be used to calculate how many more participants need to be invited to fill the quota. This procedure is further described in Fig 2, and in more detail in the supporting information (see S2 Text). Individuals of the first wave will have approximately one month overall to participate in the survey, individuals of the second wave will have two weeks to participate. Both groups will receive a reminder email one week after invitation. Additionally, the survey will be advertised on social media (Facebook, Twitter), and via student and researcher specific mailing lists, which will be seen as additional sources of participants.

Fig 2. Survey timeline and corresponding actions.

The arrow illustrates the timeline while the boxes contain details about the sampling procedure.

If the optimal sample size is achieved before the end of the set time frame (approximately one month), the survey will still be accessible to participants that were already invited, but recruitment will be discontinued. Data collection will be stopped after this time frame even if the optimal N cannot be reached.

The collection of email addresses has started on January 4th, 2021. Emails will be sent once our submission receives an in-principle-acceptance and 2960 email addresses have been collected using the procedure described above.

Participants who completed the survey will have the possibility to participate in a lottery for 40 gift cards worth 50 € each. The survey was approved by the ethics committee of Trier University, Germany.

Exclusion and missing data.

All participants that indicate that their research or studies do not fall within the scope of psychology or do not have at least a bachelor’s degree in psychology, and thus, cannot be assigned to the respective quota, will be screened out at the beginning of the survey. Thus, they will be directed to an exit page rather than to the main body of the survey, will not be counted into the quotas, and will not be considered for data analyses as the survey targets a sample from research-oriented psychology.

At the end of the survey, participants will be asked whether they responded faithfully. Only data of participants that indicated having at least a bachelor’s degree in psychology, indicated faithful participation, and completed all pages will be counted for quota fulfillment. By screening the open text inputs, it will additionally be checked if participants included any inappropriate responses like advertising or offensive comments. Those participants will be excluded from analyses. Data from all participants who indicated faithful participation, who were not excluded based on inappropriate answers, and completed all pages of the survey will be analyzed when testing the planned hypotheses (see section hypotheses tests). Additionally, responses from participants who were not screened out or excluded based on inappropriate responses, and completed the survey only partially, will be used for calculating descriptive statistics where applicable (see section descriptive reports). A sensitivity analysis will be performed to examine whether and to what extent the exclusion of incomplete datasets for the hypotheses tests affects the results.


The online survey was created with the software SoSci Survey (version 3.2.29) [54] and will be supplied via It will be presented in English. Items of the survey can be categorized in nine categories: 1) Sociodemographic questions, 2) items concerning the general usage of preregistration, 3) an attitude scale consisting of 24 items indicating the overall attitudes of participants regarding the concept of preregistration, 4) a subjective norm scale of eight items representing perceived social norms and pressure, 5) a perceived behavioral control scale of five items inquiring about researchers’ perceived control over the potential preregistration of their studies, 6) an intention scale of three items inquiring about researchers’ intention to use preregistration in the future, 7) items about motivations (a motivation scale including ten items measuring how strongly participants agree with potential motivations to preregister, plus additional more open items), 8) items about perceived obstacles (an obstacle scale featuring ten items measuring how strongly participants agree with potential obstacles to preregister, plus additional more open items), and 9) various open questions inquiring about suggestions for improving preregistration. Some items were adapted from other surveys that focused on related topics [25,35] and are complemented with additional items generated in reference to other theoretical and empirical works on preregistration, open science [31,51,53,55], and the theory of planned behavior [37,38,46,56]. These sources were selected based on their theoretical relevance to the topic at hand (i.e., open science practices and theory of planned behavior). The identified items were either adapted (i.e., wording was changed slightly to make them easier to understand or make them fit the present topic, e.g., “I have more trust in research findings when the study has been preregistered.” instead of “… when the respective data are published.” [35]) or new items were generated using the concepts presented in the literature based on face validity. The items of the present survey as well as the original items including references are available in the supporting information (see S1 Table). For most of the original items, no validity measures are given in the literature. Only for the theory of planned behavior, predictive validity is considered in more detail [46]. However, since none of the survey items were used in their original form but were adapted or newly created based on the given concepts, no validity measures can be provided. Instead, only face validity is assumed. This validity evaluation is based on Flake and Fried [57].

Different item formats will be included in the survey. All scale items (attitudes, subjective norm, perceived behavioral control, intention, motivations, and obstacles) will be answered with a seven-point labeled answer scale as recommended by Ajzen [37] (1 = “Strongly disagree”, 2 = “Disagree”, 3 = “Slightly disagree”, 4 = “Neither agree nor disagree”, 5 = “Slightly agree”, 6 = “Agree”, 7 = “Strongly agree”). Scales will be recoded from “1 to 7” to “-3 to +3” for data analyses yielding a middle category which has absolute meaning. For the statistical analyses, the mean scores of the scales will be used to measure how participants perceive 1) preregistration (attitude scale), 2) subjective norms regarding preregistration (subjective norm scale), 3) their own control about using preregistration or not (behavioral control scale), 4) their intention to use preregistration in the future (intention scale), and 5) their motivations (motivation scale) and 6) obstacles to preregister (obstacle scale). For each participant, the mean for each scale will be calculated and used as the score.

Other items will include a single or multiple choice response format, or the option for open text input. Whenever applicable, options are displayed in randomized order to eliminate potential sequence effects.


Participants will receive the link to the survey via personal email or social media call. After the welcoming page, participation information will be displayed and informed consent needs to be provided to proceed. Furthermore, a captcha (arithmetic task) needs to be completed as a safeguard against bot responses. Then, participants will complete the main body of the survey which successively focuses on the different item categories (sociodemographic questions, general usage of preregistration, attitudes, subjective norms, perceived behavioral control, intention, motivations, perceived obstacles, and suggestions for improvement). Before any items related to preregistration are shown, a definition of preregistration is presented and correct understanding is checked, to ensure that all participants answer the items with the same concept in mind. At the end of the survey, participants will have the option to participate in the lottery and to sign up for receiving a preprint of the survey results by entering their email address, which will be saved separately from the other data. Lastly, a debriefing will follow that states the aim of the survey and the research questions that are investigated. Overall, completing the survey will take about 20–25 minutes. Screen recordings of the survey’s procedure are available in the supporting information (see S1 and S2 Videos).

Data analysis

This Registered Report Protocol was written in R Markdown (version 2.7) [58,59]. R (version 4.0.5) [60] and the R-packages afex (version 0.27–2) [61], Amelia (version 1.7.6) [62], beeswarm (version 0.2.3) [63], biotools (version 3.1) [64], car (version 3.0–8) [65,66], carData (version 3.0–4) [66], corrplot2017 (version 0.84) [67], dplyr (version 1.0.0) [68], emmeans (version 1.4.7) [69], forcats (version 0.5.0) [70], ggplot2 (version 3.3.2) [71], gplots (version 3.0.3) [72], heplots (version 1.3–5) [73,74], lattice (version 0.20–41) [75], lm.beta (version 1.5–1) [76], lme4 (version 1.1–23) [77], MASS (version 7.3–53.1) [78], Matrix (version 1.3–2) [79], MBESS (version 4.7.0) [80], mlbench (version 2.1–1) [81], MVN (version 5.8) [82], olsrr (version 0.5.3) [83], papaja (version [84], popbio (version 2.7) [85], psych (version [86], purrr (version 0.3.4) [87], Rcpp (version [88,89], readr (version 1.3.1) [90], reshape2 (version 1.4.4) [91], rpanel (version 1.1–4) [92,93], sp (version 1.4–2) [94], SpatialEpi (version 1.2.3) [95], stringr (version 1.4.0) [96], tibble (version 3.0.1) [97], tidyr (version 1.1.0) [98], tidyverse (version 1.3.0) [99], tkrplot (version 0.0–24) [100], and writexl (version 1.3) [101] will be used for the analyses. Once the study has been conducted, StatTag (version 6.0.1) [102] will be used to insert the results calculated in R into the manuscript. Scripts for preprocessing, assumption testing and all analyses will be available in the supporting information (see S1S3 Scripts). Additionally, all data (including meta-data about variables and values) will be publicly accessible after study completion in PsychArchives (, a platform for sharing psychological data sets and other digital research materials.


Data will be processed in the following way: Responses from all scale items will be recoded as described in the section material, and polarity of negatively poled items will be reversed. Additionally, multiple choice questions will be recoded (originally: 1 = “not checked” and 2 = “checked”; new: 0 = “not checked” and 1 = “checked”), and single choice items will be transformed into factors. Comments will be inspected. Furthermore, data will be screened for any inappropriate responses (i.e., advertising, offenses) which would lead to an overall exclusion.

As a quality check, the attitude, subjective norm, perceived behavioral control, intention, motivation, and obstacle items will be inspected with respect to floor or ceiling effects which will then be excluded from the analyses (specifically, items for which ≥ 90% of participants indicated the lowest or highest category). Additionally, Cronbach’s alpha will be calculated as an indicator for reliability for each scale, and items for which Cronbach’s alpha of the scale is at least .10 higher if they are dropped will be excluded. The remaining items will be used to calculate mean scores for the attitude, subjective norm, perceived behavioral control, intention, motivation, and obstacle scale respectively for each participant. For this quality check, only complete datasets will be considered. All preprocessing scripts are available in the supporting information (see S1 Script).

Descriptive reports.

To describe the general behavior of participants when filling out the survey, we will calculate the response rate, the number of screened out participants, the number of complete and incomplete survey participations, the completion time (M, SD, range), and the database from which participants were recruited.

We will further describe the composition of the obtained sample by providing frequency tables or plots and descriptive statistics (M, SD, range) for selected sociodemographic information (e.g., gender, age, country of residence, academic degree, main research topic, number of years worked in research) separately for participants who preregistered in the past vs. participants without preregistration experience. Means, standard deviations, and ranges for perceived importance of preregistration and intention to preregister in the near future will be calculated. We will furthermore compare these between researchers with different academic degrees. If the respective group means do not differ as determined by a one-factorial ANOVA, we will report the overall mean, standard deviation and range across groups.

To gain insight into how common preregistration is in the general population of psychological researchers, we will report the mean proportion of participants who used preregistration in the past and how many preregistrations they created on average, across the sample obtained from the article databases, that is, excluding participants recruited from OSF Registries.

Moreover, also including the sample recruited from OSF Registries, participants’ responses to the following questions will be listed as frequency tables or in the text: 1) Where did they learn about preregistration, 2) what motivated them to create their first preregistration (only addressing participants with preregistration experience), 3) what problems did researchers experience who already used preregistration, and in turn, what worries do researchers report who have not preregistered yet (and do these match?), 4) how often do participants read other researchers’ preregistrations, and 5) by whom is the researchers’ decision to preregister influenced the most, which is compared between researchers with different academic degrees. Wherever applicable, the results will be compared between participants that have preregistered before or not.

Open text inputs of both the open “other” options of selection items and the open text input items (i.e., what are benefits and drawbacks of preregistration, what might be positive and negative consequences of mandatory preregistration, what are reasons against preregistration, how and why did participants’ motivation to preregister change, and what are suggestions for increasing the motivation, decreasing obstacles, and for other improvements regarding preregistration) will be analyzed to identify common themes. The following mixed-methods approach will be used: Two to four coders will qualitatively identify themes mentioned by the participants and subsequently categorize all responses accordingly in order to receive a frequency measure that will be analyzed quantitatively. Specifically, open text inputs of each item will be shuffled and the first 10% of these shuffled responses will be used to establish initial categories of themes. Coders will read the responses and add each theme as a column in a coding sheet. It will be coded whether the respective theme appears in the other responses (= 1) or not (= 0). If a coder encounters new relevant themes, they will be added and coded later. Nonsense responses will be excluded during coding. Ambiguities will be discussed and solved in pairs. If no solution can be found, a third coder will be consulted. After all responses have been coded, the sums for each column (= theme) will be calculated to obtain the frequency of how often a theme was mentioned over all responses. Results from the open text input items will be displayed in frequency tables, and the results of the analysis of “other” options will be added to the individual items’ response presentations (see above). Again, responses will be compared between participants with and without preregistration experience. Both the coding sheets and the open text inputs will be published alongside the other data once the study has been conducted. Furthermore, all analysis scripts will be included in the supporting information (see S3 Script).

Hypotheses tests.

In order to find out how the intention to preregister is formed, and whether research experience has an impact on attitudes and the intensity of motivations and perceived obstacles regarding preregistration, we will analyze participants’ responses to the attitude, subjective norm, perceived behavioral control, intention, motivation, and obstacle scale (see S3 Script for all analyses). Only complete datasets (see section exclusion and missing data) will be considered for these analyses. Before the analyses, the aforementioned scales will be inspected, and the mean, standard deviation, and distribution of responses will be displayed per item in a table for easy inspection. Next, means, standard deviations, and ranges will be computed for each scale.

A significance level of α = .05 will be used for our hypotheses tests. Before conducting statistical analyses, assumptions will be tested for each method (see S2 Script). If assumptions are violated and tests are not robust against these violations, alternative methods will be used as described below.

For hypothesis 1, we predict that attitudes, subjective norms, and perceived behavioral control influence researchers’ intention to preregister their studies in the near future. Perceived importance of the topic of preregistration will be included as moderator for attitudes, and previous preregistration experience will serve as a control variable. To test this hypothesis, a moderated multiple regression model will be computed. Assumptions for this model include linearity, uncorrelated predictors, independence and normality of residuals, and homogeneity of variance. Linearity and uncorrelated predictors are assumed on a theoretical basis. Yet, there might be a high multicollinearity for the product term (attitudes x perceived importance). This will be examined by calculating the tolerance statistics as well as the variance inflation factor. If the multicollinearity is too high for these two predictors (variance inflation factor > 10), importance will be aggregated with the attitudes scale instead of including it as a moderator, since it might be just another aspect of attitudes. In this case, the item concerning perceived importance would be included as another item in the attitude scale which would now include 25 instead of 24 items. The mean score would then be re-calculated and the new score would be included in the analyses. The independence of residuals is furthermore achieved by the survey’s design. Normality of residuals and homogeneity of variance will be investigated by plotting the residuals (scatterplots, Q-Q plots and histograms). A violation of the normality of residuals is not impactful with a high N, as it is the case with our sample, thus a violation would not be problematic. If the assumption of homogeneity of variance is violated, weighted least squares regression will be used instead of ordinary least squares.

After testing these assumptions, the regression model will be computed: The intention (i.e., mean score of three items) will be included as dependent variable. As predictors, the attitude scale (i.e., mean score of 24 items inquiring about how positive preregistrations are perceived), perceived importance of preregistration (one item), and their product term, the subjective norm scale (i.e., mean score of eight items inquiring about how beneficial participants perceive the social norm of using preregistration), the perceived behavioral control scale (i.e., mean score of five items inquiring about perceived controllability of the behavior), and previous preregistration experience (one item: yes vs. no) will be included. We expect that higher scores on the attitude scale (i.e., more positive attitudes, see hypothesis 1.1), higher scores on the subjective norm scale (i.e., higher perceived social pressure, see hypothesis 1.4), higher scores on the perceived behavioral control scale (i.e., higher perceived control, see hypothesis 1.5) are positive predictors for participants’ intention scores (i.e., higher intention to preregister their studies in the near future). Additionally, we expect that perceived importance of preregistration is a positive predictor for intention (see hypothesis 1.3), and also significantly moderates the strength of the influence of attitudes (see hypothesis 1.2). Furthermore, we expect that the overall model including all predictors can significantly predict researchers’ intention to preregister their studies in the future (see hypothesis 1.6). Because of these directional hypotheses, the regression weights will be tested in a one-tailed fashion. Standardized regression weights will be interpreted and compared. Preregistration experience will be included into the model to control for a potential sampling bias. We have no strong predictions regarding the direction of this effect, since it might be the case that a) researchers who have used preregistration before are more likely to preregister again or b) that their preregistration experiences were negative and they will less likely preregister again. Therefore, in contrast to the other predictors, this effect will be tested in a two-tailed fashion.

If all main predictors (attitudes, subjective norm, perceived behavioral control, perceived importance, and its interaction with attitudes) and the overall model are significant (support for all hypotheses 1.11.6), we will conclude that the theory of planned behavior is indeed well suited to be used to predict researchers’ intention to preregister. If the overall model, but not all predictors are significant (support for hypothesis 1.6, no or partly support of hypotheses 1.11.5), the theory of planned behavior will still be seen as applicable, with the significant predictors being more relevant in the context of preregistration than the non-significant ones. If the individual predictors as well as the overall model are not significant (no support for hypotheses 1.11.6), it will be concluded that the theory of planned behavior cannot be applied to the context of predicting researchers’ intention to preregister, but that it might be beneficial to test it again with other open science techniques.

The second hypothesis predicts that research experience (i.e., the years someone has worked in research) influences researchers’ attitudes (see hypothesis 2.1) as well as motivations (see hypothesis 2.2) and the perception of obstacles (see hypothesis 2.3). To investigate this, three multiple regression models will be computed: Mean scores on the attitude scale (i.e., mean score of 24 items inquiring about how positive preregistrations are perceived), motivation scale (i.e., mean score of ten items measuring how strongly participants agree with potential motivations to preregister), and obstacle scale (i.e., mean score of ten items measuring how strongly participants agree with potential obstacles to preregister) will serve as dependent variables in the respective regression models, while research experience (i.e., the years someone has worked in research) and the preregistration experience (control variable) will serve as predictors. As our hypotheses regarding the impact of research experience on attitudes, and intensity of motivations and perceived obstacles are non-directional, the regression weights will be analyzed with two-tailed tests. As three regressions will be conducted to test one set of hypotheses, a Bonferroni-Holm correction will be used to adjust the alpha level. The assumptions for these regressions will be tested in the same manner as those of the regression model for hypothesis 1 (see above).

If all three regression models show significant effects (support for hypotheses 2.12.3), we will conclude that research experience has an influence on how researchers perceive preregistration, as well as its motivations and obstacles. If only one or two of the regression models are significant (partly support of hypotheses 2.12.3), we will conclude that only for these scales, research experience has an influence, while if no significant effects can be detected (no support for hypotheses 2.1–2.3), we will conclude that, against our predictions, research experience has no influence on attitudes, motivations, and perceived obstacles concerning preregistration.

Pilot study

A pilot study was conducted to estimate the response rate and to test the recruitment method as well as the survey items. The pilot study featured mostly the same items as the main study but additional items, prompting participants for feedback about comprehensibility via open text input fields, were presented (“Did you have any problems answering the items of this page? Was anything unclear?”). Some items were adjusted based on the results. Quotas were not applied.

We invited N = 200 participants to partake in the pilot study (100 were invited from OSF Registries, 50 from Web of Science, and 50 from PubMed) and sent a reminder one week after the initial invitation. Participations typically occurred shortly after each email contact. The pilot study was accessible for one month overall. In this time, 29 participants (17 PhD students, three postdocs, seven professors, and two members of other academic groups which were screened out) started the survey of which 20 completed it (14 PhD students and 6 professors), yielding an overall response rate of 10%. Out of these 20, 18 (90%) participants had preregistration experience. No specific patterns in dropout behavior were found (e.g., dropping out at a certain position). No floor or ceiling effects were found. Reliability analyses for the measured scales were conducted, which showed a very high reliability for the attitude (α = 0.90) and obstacle scale (α = 0.87), an acceptable reliability for the motivation scale (α = 0.77), and a moderate reliability for the subjective norm (α = 0.65) and perceived behavioral control scale (α = 0.63). One item of the motivation scale correlated negatively with the remaining scale items and Cronbach’s alpha increased by removing it (item M2, see S1 Table). Nevertheless, the item will remain in the survey and will be checked again in the main study, since the pilot results are based on a very small sample and can only be interpreted with caution.

After conducting the pilot study, the sampling strategy was revised by extending the recruitment of participants to four instead of two general databases for psychological articles and by inviting less participants via OSF Registries (10% instead of 50%). Moreover, some comments indicated ambiguous interpretations which led us to slightly modify the wording of respective items or to change categories (i.e., academic position was replaced by academic degree in the sociodemographic part of the survey). Lastly, a few items were added, e.g., two items assessing intention following a manual about constructing questionnaires based on the theory of planned behavior [56].

Supporting information

S1 Text. Power analyses.

Power analyses for both hypotheses are presented here. For hypothesis 1, power analyses were performed for the entire model as well as for the individual predictors.


S2 Text. Data collection procedure.

The data collection procedure that is displayed in Fig 2 is further specified. Particularly, the specific procedures for collecting contact addresses as well as for inviting participants are described.


S1 Table. Survey items.

The items of the survey are presented. If they were derived from other studies, the original items as well as references are given.


S1 Video. Screen recording of the survey procedure (version A).

This screen recording shows the procedure of the survey for participants that responded with “yes” on the item “Have you preregistered a study before?” (as this is a filter question for some of the following items) and who answered the knowledge check of the preregistration definition correctly.


S2 Video. Screen recording of the survey procedure (version B).

This screen recording shows the procedure of the survey for participants that responded with “no” on the item “Have you preregistered a study before?” (as this is a filter question for some of the following items) and who were shown the definition of preregistration again due to errors in the knowledge check.


S1 Script. Preprocessing script (preliminary).

This R script contains all steps of preprocessing. The hereby processed data will be used for the assumption tests as well as for the descriptive reports and hypotheses tests. If necessary, small changes may be made to enable correct functioning (when variables have been inserted).


S2 Script. Assumption testing script (preliminary).

This R script will be used for testing the assumptions of the statistical models used to test hypotheses 1 and 2. If necessary, small changes may be made to enable correct functioning (when variables have been inserted).


S3 Script. Analysis script (preliminary).

This script will be used to conduct the descriptive reports as well as the hypotheses tests of the survey. If necessary, small changes may be made to enable correct functioning (when variables have been inserted). Furthermore, analyses might be added if alternative approaches need to be used (due to assumption violation).



We are grateful to H. Bargon and A. Kuznik for their help with testing the survey, and their future help with recruitment and coding, and also to T. Dauber and J. Pauquet for testing the survey and proofreading.


  1. 1. Ioannidis JPA. Why most published research findings are false. PLoS Med. 2005;2: e124. pmid:16060722
  2. 2. Open Science Collaboration. Estimating the reproducibility of psychological science. Science. 2015;349: aac4716. pmid:26315443
  3. 3. Shrout PE, Rodgers JL. Psychology, Science, and Knowledge Construction: Broadening Perspectives from the Replication Crisis. Annu Rev Psychol. 2018;69: 487–510. pmid:29300688
  4. 4. Stevens JR. Replicability and Reproducibility in Comparative Psychology. Front Psychol 2017;8: 862. pmid:28603511
  5. 5. Baker M. 1,500 scientists lift the lid on reproducibility. Nature. 2016;533: 452–454. pmid:27225100
  6. 6. Klein RA, Ratliff KA, Vianello M, Adams RB, Bahník Š, Bernstein MJ, et al. Investigating Variation in Replicability. Soc Psychol (Gott). 2014;45: 142–152.
  7. 7. Camerer CF, Dreber A, Holzmeister F, Ho T-H, Huber J, Johannesson M, et al. Evaluating the replicability of social science experiments in Nature and Science between 2010 and 2015. Nat Hum Behav. 2018;2: 637–644. pmid:31346273
  8. 8. Simmons JP, Nelson LD, Simonsohn U. False-positive psychology: Undisclosed flexibility in data collection and analysis allows presenting anything as significant. Psychol Sci. 2011;22: 1359–1366. pmid:22006061
  9. 9. John LK, Loewenstein G, Prelec D. Measuring the prevalence of questionable research practices with incentives for truth telling. Psychol Sci. 2012;23: 524–532. pmid:22508865
  10. 10. Munafò MR, Nosek BA, Bishop DVM, Button KS, Chambers CD, Percie du Sert N, et al. A manifesto for reproducible science. Nat Hum Behav. 2017;1: 0021. pmid:33954258
  11. 11. Kerr NL. HARKing: Hypothesizing after the results are known. Pers Soc Psychol Rev. 1998;2: 196–217. pmid:15647155
  12. 12. Franco A, Malhotra N, Simonovits G. Underreporting in Psychology Experiments. Soc Psychol Personal Sci. 2016;7: 8–12.
  13. 13. Wagenmakers E-J, Wetzels R, Borsboom D, van der Maas HLJ, Kievit RA. An Agenda for Purely Confirmatory Research. Perspect Psychol Sci. 2012;7: 632–638. pmid:26168122
  14. 14. de Vries YA, Roest AM, de Jonge P, Cuijpers P, Munafò MR, Bastiaansen JA. The cumulative effect of reporting and citation biases on the apparent efficacy of treatments: the case of depression. Psychol Med. 2018;48: 2453–2455. pmid:30070192
  15. 15. Nosek BA, Ebersole CR, DeHaven AC, Mellor DT. The preregistration revolution. Proc Natl Acad Sci U S A. 2018;115: 2600–2606. pmid:29531091
  16. 16. van ‘t Veer AE, Giner-Sorolla R. Pre-registration in social psychology—A discussion and suggested template. J Exp Soc Psychol. 2016;67: 2–12.
  17. 17. Kaplan RM, Irvin VL. Likelihood of Null Effects of Large NHLBI Clinical Trials Has Increased over Time. PLoS One. 2015;10: e0132382. pmid:26244868
  18. 18. Swaen GG, Teggeler O, van Amelsvoort LG. False positive outcomes and design characteristics in occupational cancer epidemiology studies. Int J Epidemiol. 2001;30: 948–954. pmid:11689501
  19. 19. De Angelis C, Drazen JM, Frizelle FA, Haug C, Hoey J, Horton R, et al. Clinical trial registration: A statement from the International Committee of Medical Journal Editors. CMAJ. 2004;171: 606–607. pmid:15367465
  20. 20. Nosek BA, Beck ED, Campbell L, Flake JK, Hardwicke TE, Mellor DT, et al. Preregistration Is Hard, And Worthwhile. Trends Cogn Sci. 2019;23: 815–818. pmid:31421987
  21. 21. Nosek BA, Lindsay DS. Preregistration Becoming the Norm in Psychological Science. 2018 Feb 28 [cited 2021 June 7]. In: APS Obs [Internet]. Available from:
  22. 22. Wagenmakers E-J, Dutilh G. Seven Selfish Reasons for Preregistration. 2016 Oct 31 [cited 2021 June 7]. In: APS Obs [Internet]. Available from: pmid:28736493
  23. 23. Christensen G, Wang Z, Levy Paluck E, Swanson N, Birke DJ et al. Open Science Practices are on the Rise: The State of Social Science (3S) Survey. MetaArXiv [Preprint]. 2019 [posted 2019 Oct 18; revised 2019 Dec 18; cited 2021 June 7]. Available from:
  24. 24. Hardwicke TE, Thibault RT, Kosie JE, Wallach JD, Kidwell M, Ioannidis J. Estimating the prevalence of transparency and reproducibility-related research practices in psychology (2014–2017). MetaArXiv [Preprint]. 2020 [posted 2020 Jan 2; revised 2020 June 17; cited 2021 June 7]. Available from:
  25. 25. Stürmer S, Oeberst A, Trötschel R, Decker O. Early-Career Researchers’ Perceptions of the Prevalence of Questionable Research Practices, Potential Causes, and Open Science. Soc Psychol (Gott). 2017;48: 365–371.
  26. 26. Goldin-Meadow S. Why Preregistration Makes Me Nervous. 2016 Aug 31 [cited 2021 June 7]. In: APS Obs [Internet]. Available from: pmid:28736493
  27. 27. Nosek BA. Preregistration: Improve Research Rigor, Reduce Bias [Video]. 2018. Available from:
  28. 28. Szollosi A, Donkin C. Arrested theory development: The misguided distinction between exploratory and confirmatory research. PsyArXiv [Preprint]. 2019 [posted 2019 Sep 21; revised 2020 Sep 19; cited 2021 June 7]. Available from:
  29. 29. Szollosi A, Donkin C. Neglected Sources of Flexibility in Psychological Theories: From Replicability to Good Explanations. Comput Brain Behav. 2019;2: 190–192.
  30. 30. Szollosi A, Kellen D, Navarro DJ, Shiffrin R, van Rooij I, van Zandt T, et al. Is Preregistration Worthwhile? Trends Cogn Sci. 2020;24: 94–95. pmid:31892461
  31. 31. Ikeda A, Xu H, Fuji N, Zhu S, Yamada Y. Questionable research practices following pre-registration. Japanese Psychological Review. 2019;62: 281–295. Available from:
  32. 32. Claesen A, Gomes SLBT, Tuerlinckx F, Vanpaemel W. Preregistration: Comparing Dream to Reality. PsyArXiv [Preprint]. 2019 [posted 2019 May 9; revised 2020 June 25; cited 2021 June 7]. Available from:
  33. 33. Ofosu G, Posner DN. Pre-analysis Plans: A Stocktaking. MetaArXiv [Preprint]. 2019 [posted 2019 Dec 13; revised 2019 Dec 14; cited 2021 June 7]. Available from:
  34. 34. Li G, Abbade LPF, Nwosu I, Jin Y, Leenus A, Maaz M, et al. A systematic review of comparisons between protocols or registrations and full reports in primary biomedical research. BMC Med Res Methodol. 2018;18: 9. pmid:29325533.
  35. 35. Abele-Brehm AE, Gollwitzer M, Steinberg U, Schönbrodt FD. Attitudes Toward Open Science and Public Data Sharing. Soc Psychol (Gott). 2019;50: 252–260.
  36. 36. DeHaven AC, Graf C, Mellor DT, Morris E, Moylan E, Pedder S, et al. Registered Reports: Views from editors, reviewers and authors. PsyArXiv [Preprint]. 2019 [posted 2019 Sep 17; revised 2019 Sep 17; cited 2021 June 7]. Available from:
  37. 37. Ajzen I. The theory of planned behavior. Organ Behav Hum Decis Process. 1991;50: 179–211. (91)90020-T.
  38. 38. Ajzen I, Fishbein M, Lohmann , Albarracín D. The influence of attitudes on behavior. In: Albarracín D, Johnson BT, editors. The handbook of attitudes: Volume 1: Basic principles (2nd Ed). New York: Routledge; 2019. pp. 197–255.
  39. 39. Miller ZD. The Enduring Use of the Theory of Planned Behavior. Human Dimensions of Wildlife. 2017;22: 583–590.
  40. 40. Armitage CJ, Conner M. Efficacy of the Theory of Planned Behaviour: A meta-analytic review. Br J Soc Psychol. 2001;40: 471–499. pmid:11795063
  41. 41. Notani A. Moderators of Perceived Behavioral Control’s Predictiveness in the Theory of Planned Behavior: A Meta-Analysis. J Consum Psychol. 1998;7: 247–271.
  42. 42. Rivis A, Sheeran P. Descriptive norms as an additional predictor in the theory of planned behaviour: A meta-analysis. Curr Psychol. 2003;22: 218–233.
  43. 43. McEachan RRC, Conner M, Taylor NJ, Lawton RJ. Prospective prediction of health-related behaviours with the Theory of Planned Behaviour: a meta-analysis. Health Psychol Rev 2011;5: 97–144.
  44. 44. Godin G, Kok G. The theory of planned behavior: A review of its applications to health-related behaviors. Am J Health Promot. 1996;11: 87–98. pmid:10163601
  45. 45. Downs DS, Hausenblas HA. The Theories of Reasoned Action and Planned Behavior Applied to Exercise: A Meta-analytic Update. J Phys Act Health. 2005;2: 76–97.
  46. 46. Ajzen I. The theory of planned behaviour: Reactions and reflections. Psychol Health. 2011;26: 1113–1127. pmid:21929476
  47. 47. Schuman H, Presser S. Questions and answers in attitude surveys: Experiments on question form, wording, and context. New York: Academic Press; 1981.
  48. 48. Chambers C. The seven deadly sins of psychology: A manifesto for reforming the culture of scientific practice. 1st paperback ed. Princeton, New Jersey; Woodstock, Oxfordshire: Princeton University Press; 2019.
  49. 49. Faul F, Erdfelder E, Lang A-G, Buchner A. G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behav Res Methods. 2007;39: 175–191. pmid:17695343
  50. 50. Faul F, Erdfelder E, Buchner A, Lang A-G. Statistical power analyses using G*Power 3.1: Tests for correlation and regression analyses. Behav Res Methods. 2009;41: 1149–1160. pmid:19897823
  51. 51. Field SM, Wagenmakers E-J, Kiers HAL, Hoekstra R, Ernst AF, van Ravenzwaaij D. The effect of preregistration on trust in empirical research findings: Results of a registered report. R Soc Open Sci. 2020;7: 181351. pmid:32431853
  52. 52. van den Akker O, Alvarez LD, Bakker M, Wicherts JM, van Assen MALM. How do academics assess the results of multiple experiments? PsyArXiv [Preprint]. 2018 [posted 2018 Dec 16; revised 2020 June 18; cited 2021 June 7]. Available from:
  53. 53. Mathieu S, Chan A-W, Ravaud P. Use of trial register information during the peer review process. PLoS One. 2013;8: e59910. pmid:23593154
  54. 54. Leiner DJ. SoSci Survey (Version 3.2.29) [Computer software]. 2019. Available from:
  55. 55. Allen C, Mehler DMA. Open science challenges, benefits and tips in early career and beyond. PLoS Biol. 2019;17: e3000246. pmid:31042704
  56. 56. Francis J, Eccles MP, Johnston M, Walker AE, Grimshaw JM, Foy R, et al. Constructing questionnaires based on the theory of planned behaviour: A manual for health services researchers. Newcastle upon Tyne, UK: Centre for Health Services Research, University of Newcastle upon Tyne; 2004. Available from:
  57. 57. Flake JK, Fried EI. Measurement Schmeasurement: Questionable Measurement Practices and How to Avoid Them. PsyArXiv [Preprint]. 2019 [posted 2019 Jan 17; revised. 2019 Jan 18; revised 2019 Jan 19; revised 2019 Jan 22; revised 2019 Nov 2; revised 2020 Jun 18; cited 2021 June 7]. Available from:
  58. 58. Xie Y, Allaire JJ, Grolemund G. R Markdown: The Definitive Guide. Boca Raton, Florida: Chapman and Hall/CRC; 2018. Available from:
  59. 59. Allaire JJ, Xie Y, McPherson J, Luraschi J, Ushey K, Atkins A, et al. rmarkdown: Dynamic Documents for R. Version 2.7 [software]. 2020 [cited 2021 June 7]. Available from:
  60. 60. R Core Team. R: A language and environment for statistical computing. Vienna, Austria: R Foundation for Statistical Computing; 2020. Available from:
  61. 61. Singmann H, Bolker B, Westfall J, Aust F, Ben-Shachar MS. Afex: Analysis of factorial experiments. Version 0.27–2 [software]. 2020 [cited 2021 June 7]. Available from:
  62. 62. Honaker J, King G, Blackwell M. Amelia II: A program for missing data. J Stat Softw. 2011;45: 1–47.
  63. 63. Eklund A. Beeswarm: The bee swarm plot, an alternative to stripchart. Version 0.2.3 [software]. 2016 [cited 2021 June 7]. Available from:
  64. 64. Silva AR da, Malafaia G, de Menezes IPP. Biotools: An r function to predict spatial gene diversity via an individual-based approach. Genet Mol Res. 2017;16: gmr16029655. pmid:28407196
  65. 65. Fox J, Weisberg S. An R companion to applied regression. 3rd ed. Thousand Oaks CA: Sage; 2019. Available from:
  66. 66. Fox J, Weisberg S, Price B. CarData: Companion to applied regression data sets. Version 3.0–4 [software]. 2019 [cited 2021 June 7]. Available from:
  67. 67. Wei T, Simko V. R package "corrplot": Visualization of a correlation matrix. Version 0.84 [software]. 2017 [cited 2021 June 7]. Available from:
  68. 68. Wickham H, François R, Henry L, Müller K. Dplyr: A grammar of data manipulation. Version 1.0.0 [software]. 2020 [cited 2021 June 7]. Available from:
  69. 69. Lenth R. Emmeans: Estimated marginal means, aka least-squares means. Version 1.4.7 [software]. 2020 [cited 2021 June 7]. Available from:
  70. 70. Wickham H. Forcats: Tools for working with categorical variables (factors). Version 0.5.0 [software]. 2020 [cited 2021 June 7]. Available from:
  71. 71. Wickham H. Ggplot2: Elegant graphics for data analysis. New York: Springer; 2016. Available from:
  72. 72. Warnes GR, Bolker B, Bonebakker L, Gentleman R, Huber W, Liaw A, et al. Gplots: Various r programming tools for plotting data. Version 3.0.3 [software]. 2020 [cited 2021 June 7]. Available from:
  73. 73. Friendly M. HE plots for multivariate linear models. J Comput Graph Stat. 2007;16: 421–444.
  74. 74. Friendly M. HE plots for repeated measures designs. J Stat Softw. 2010;37: 1–40.
  75. 75. Sarkar D. Lattice: Multivariate data visualization with r. New York: Springer; 2008. Available from:
  76. 76. Behrendt S. Lm.beta: Add standardized regression coefficients to lm-objects. Version 1.5–1 [software]. 2014 [cited 2021 June 7]. Available from:
  77. 77. Bates D, Mächler M, Bolker B, Walker S. Fitting linear mixed-effects models using lme4. J Stat Softw. 2015;67: 1–48.
  78. 78. Venables WN, Ripley BD. Modern applied statistics with s. 4th ed. New York: Springer; 2002. Available from:
  79. 79. Bates D, Maechler M. Matrix: Sparse and dense matrix classes and methods. Version 1.3–2 [software]. 2019 [cited 2021 June 7]. Available from:
  80. 80. Kelley K. MBESS: The mbess r package. Version 4.7.0 [software]. 2020 [cited 2021 June 7]. Available from:
  81. 81. Leisch F, Dimitriadou E. Mlbench: Machine learning benchmark problems. Version 2.1–1 [software]. 2010 [cited 2021 June 7]. Available from:
  82. 82. Korkmaz S, Goksuluk D, Zararsiz G. MVN: An r package for assessing multivariate normality. R J. 2014;6: 151–162. Available from:
  83. 83. Hebbali A. Olsrr: Tools for building ols regression models. Version 0.5.3 [software]. 2020 [cited 2021 June 7]. Available from:
  84. 84. Aust F, Barth M. papaja: Create APA manuscripts with R Markdown. Version [software]. 2020 [cited 2021 June 7]. Available from:
  85. 85. Stubben CJ, Milligan BG. Estimating and analyzing demographic models using the popbio package in r. J Stat Softw. 2007;22.
  86. 86. Revelle W. Psych: Procedures for psychological, psychometric, and personality research. Version [software]. 2019 [cited 2021 June 7]. Available from:
  87. 87. Henry L, Wickham H. Purrr: Functional programming tools. Version 0.3.4 [software]. 2019 [cited 2021 June 7]. Available from:
  88. 88. Eddelbuettel D, François R. Rcpp: Seamless R and C++ integration. J Stat Softw. 2011;40: 1–18.
  89. 89. Eddelbuettel D, Balamuta JJ. Extending extitR with extitC++: A Brief Introduction to extitRcpp. PeerJ Preprints [Preprint]. 2017;5: e3188v1 [posted 2017 Aug 25; cited 2021 June 7]. Available from:
  90. 90. Wickham H, Hester J, Francois R. Readr: Read rectangular text data. Version 1.3.1 [software]. 2018 [cited 2021 June 7]. Available from:
  91. 91. Wickham H. Reshaping data with the reshape package. J Stat Softw. 2007;21: 1–20.
  92. 92. Bowman A, Crawford E, Alexander G, Bowman RW. rpanel: Simple interactive controls for R functions using the tcltk package. J Stat Softw. 2007;17: 1–18.
  93. 93. Bowman AW, Gibson I, Scott EM, Crawford E. Interactive teaching tools for spatial sampling. J Stat Softw. 2010;36: 1–17.
  94. 94. Pebesma EJ, Bivand RS. Classes and methods for spatial data in R. R News. 2005;5: 9–13. Available from:
  95. 95. Kim AY, Wakefield J. SpatialEpi: Methods and data for spatial epidemiology. Version 1.2.3 [software]. 2018 [cited 2021 June 7]. Available from:
  96. 96. Wickham H. Stringr: Simple, consistent wrappers for common string operations. Version 1.4.0 [software]. 2019 [cited 2021 June 7]. Available from:
  97. 97. Müller K, Wickham H. Tibble: Simple data frames. Version 3.0.1 [software]. 2020 [cited 2021 June 7]. Available from:
  98. 98. Wickham H, Henry L. Tidyr: Tidy messy data. Version 1.1.0 [software]. 2020 [cited 2021 June 7]. Available from:
  99. 99. Wickham H, Averick M, Bryan J, Chang W, McGowan LD, François R, et al. Welcome to the tidyverse. J Open Source Softw. 2019;4: 1686.
  100. 100. Tierney L. Tkrplot: TK rplot. Version 0.0–24 [software]. 2018 [cited 2021 June 7]. Available from:
  101. 101. Ooms J. Writexl: Export data frames to excel ‘xlsx’ format. Version 1.3 [software]. 2019 [cited 2021 June 7]. Available from:
  102. 102. Welty LJ, Rasmussen LV, Baldridge AS, Whitley E. StatTag. Version 6.0.1 [software]. Chicago, Illinois, United States: Galter Health Sciences Library. 2016 [cited 2021 June 7].