Skip to main content
Browse Subject Areas

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

The Use and Abuse of Transcranial Magnetic Stimulation to Modulate Corticospinal Excitability in Humans

  • Martin E. Héroux,

    Affiliations Neuroscience Research Australia, Randwick, NSW, Australia, University of New South Wales, Randwick, NSW, Australia

  • Janet L. Taylor,

    Affiliations Neuroscience Research Australia, Randwick, NSW, Australia, University of New South Wales, Randwick, NSW, Australia

  • Simon C. Gandevia

    Affiliations Neuroscience Research Australia, Randwick, NSW, Australia, University of New South Wales, Randwick, NSW, Australia


21 Jan 2016: Héroux ME, Taylor JL, Gandevia SC (2016) Correction: The Use and Abuse of Transcranial Magnetic Stimulation to Modulate Corticospinal Excitability in Humans. PLOS ONE 11(1): e0147890. View correction


The magnitude and direction of reported physiological effects induced using transcranial magnetic stimulation (TMS) to modulate human motor cortical excitability have proven difficult to replicate routinely. We conducted an online survey on the prevalence and possible causes of these reproducibility issues. A total of 153 researchers were identified via their publications and invited to complete an anonymous internet-based survey that asked about their experience trying to reproduce published findings for various TMS protocols. The prevalence of questionable research practices known to contribute to low reproducibility was also determined. We received 47 completed surveys from researchers with an average of 16.4 published papers (95% CI 10.8–22.0) that used TMS to modulate motor cortical excitability. Respondents also had a mean of 4.0 (2.5–5.7) relevant completed studies that would never be published. Across a range of TMS protocols, 45–60% of respondents found similar results to those in the original publications; the other respondents were able to reproduce the original effects only sometimes or not at all. Only 20% of respondents used formal power calculations to determine study sample sizes. Others relied on previously published studies (25%), personal experience (24%) or flexible post-hoc criteria (41%). Approximately 44% of respondents knew researchers who engaged in questionable research practices (range 32–70%), yet only 18% admitted to engaging in them (range 6–38%). These practices included screening subjects to find those that respond in a desired way to a TMS protocol, selectively reporting results and rejecting data based on a gut feeling. In a sample of 56 published papers that were inspected, not a single questionable research practice was reported. Our survey revealed that approximately 50% of researchers are unable to reproduce published TMS effects. Researchers need to start increasing study sample size and eliminating—or at least reporting—questionable research practices in order to make the outcomes of TMS research reproducible.


Transcranial magnetic stimulation (TMS) is a popular technique in neuroscience. Its popularity stems from the allure of drawing important inferences about human brain function with a seemingly non-invasive, and certainly non-painful technique with few serious side effects [1, 2]. Originally introduced as a technique in clinical neurophysiology to assess central and peripheral motor conduction [3] and subsequently motor cortical physiology (e.g., [4]), TMS has blossomed into a method to explore brain physiology. It comes in many styles: from single pulses to complex sets of repetitive pulses, several of which have gained favour as methods to assess and modulate particular aspects of cortical function [5, 6]. These include paired associative stimulation [7] and various forms of repetitive stimulation (such as theta burst stimulation, e.g. [8]).

In our experience some of the published protocols work but they are often difficult to replicate routinely (e.g. [9]). Our colleagues frequently tell us at conferences that they also have difficulty reproducing the original published effects. These difficulties may in part reflect the small effects being investigated—usually in a small number of subjects—which results in low statistical power [10]. These issues are compounded by the presence of physiological and non-physiological factors that can further limit the reproducibility of published findings [1017]. The apparent success of these techniques may be fuelled by the bias to publish positive rather than negative results [1820], which can lead to questionable research practices [2125] and inflated effect sizes [10, 2630].

To determine whether our experience reflects that of the broader research community, we conducted on online survey to determine the prevalence of non-reproducible results in published and non-published studies that have used TMS to alter motor cortical or corticospinal excitability in humans. To gain insight into the cause of this low reproducibility, survey respondents were also questioned on various research practices known to contribute to low statistical power and exaggerated effect sizes.


To gain an overview of the field using TMS to probe or alter motor cortical and corticospinal excitability in humans, we invited the first and last authors of relevant publications to complete an anonymous internet-based survey (see S1 File) approved by the University of New South Wales Health Sciences Ethics Board. An initial search was conducted in PubMed in February 2014 for studies that had used TMS or one of its common variants to modulate motor cortical and corticospinal excitability (see S2 File). References (n = 1,486) were reviewed by one of the authors (MH) and those that were clearly not relevant were excluded. The email addresses of first and last authors of the remaining references were obtained from the manuscript or by web search. A total of 153 researchers were invited to take part in the survey. After completing the survey, researchers were entered into a draw—independently conducted by the local IT department—to win an iPad.

In brief, the survey asked about the number of years respondents worked in the field and the type of TMS protocols they had previously used. We then asked about their published and unpublished studies that involved TMS and, in particular, how study sample sizes were determined and whether the results were in line with the original published findings. Finally, we asked respondents how they thought other researchers perform and report TMS studies and, using the same questions, we asked how they themselves performed and reported TMS studies. On completing the survey, respondents were invited to provide additional comments.

A sub-sample of papers were reviewed to determine whether the questionable research practices listed in our survey [Q8] are routinely reported in the literature. To obtain a random and representative sample, papers focusing on theta-burst stimulation published between 2010-Oct 2014 were identified and digital copies obtained (see S2 File). A total of 56 papers were retained. Each paper was reviewed by one of the authors (SG) to determine whether the questionable research practices listed in our survey were reported and results were verified by a second author (MH).


Of 153 invited researchers who use TMS to modulate corticospinal excitability we received 47 completed surveys (see S3 File). Respondents, who could select more than one research area, worked in a variety of fields: neuroscience (19.1%), motor control (21.0%), neurophysiology (21.7%), clinical neurology (15.9%), rehabilitation (14.0%) and psychology (4.5%) [Q1]. They had been working in these fields for a mean of 14.2 years (95% confidence interval 12.1–16.4; range 1–30 years) [Q2] and had published a career mean number of 46.6 papers (CI 30.1–65.3) using TMS to study the human motor system [Q4]. A subset of these papers (16.4, CI 10.8–22.0) used TMS techniques specifically to alter corticospinal excitability. When asked about relevant file-drawer papers (i.e. studies that were completed but not published), all but 9 of the respondents had at least one such paper, with the mean being 4.0 papers (CI 2.5–5.7) [Q4]. Three respondents reported large numbers of file-drawer papers (10, 15 and 30 papers), which exceeded their number of published TMS-related papers.

Respondents had experience with a variety of TMS techniques, and several respondents had experience with more than one technique. Of the many forms of TMS, repetitive TMS at low frequency (<1Hz) and high frequency (>1Hz) were commonly used (19.7% and 19.1%), as was paired associative stimulation (20.1%). Slightly less frequent were intermittent (16.3%) and continuous (16.3%) theta burst stimulation [Q3].

A range of methods were used by respondents to determine sample size in TMS studies [Q5]. Respondents could select more than one method. Of the 140 responses, only 20% indicated that formal power calculations were used; a greater number indicated a reliance on previously published studies (25%) or personal experience (24%). A further 15% indicated that sample size was set prior to the start of the study, but additional subjects were tested if needed. In 5% of cases the sample size was adjusted based on how the data were looking. The remaining responses indicated that sample size was set prior to start of the study, but fewer subjects were tested because a clear effect was (5.7%) or was not (5.7%) observed.

For the various TMS protocols they had used, we attempted to gauge whether investigators were able to reproduce a similar effect to what was reported in the original studies [Q6]. The percentage of respondents who answered yes was 61% for paired associative stimulation, 45% for continuous theta burst stimulation, 45% for intermittent theta burst stimulation, 60% for low frequency (<1 Hz) and 59% for high frequency (>1 Hz) repetitive TMS. The size of the observed effects was either smaller (32 responses), larger (3 responses) or the same as those by the original studies (64 responses). The remaining respondents were able to reproduce these effects only sometimes (56 responses) or not at all (18 responses). This applied to the majority of respondents who had used either form of theta burst stimulation. Respondents who were unable to reproduce an effect similar to that which was originally published were twice as likely to stop using the TMS protocol than seek to publish the negative results (12 vs 6 reports). Also, there was no difference in respondents’ years of experience and whether they were able or unable to reproduce published results (Wilcoxon rank sum test, p = 0.441).

Finally, we asked whether respondents used practices that could increase the chance of finding statistically significant results. Response rates for these questionable practices are presented in Table 1. On average, 44% of respondents knew researchers who engaged in these practices (range 30–68%), whereas only 18% admitted to these practices (range 6–38%). Almost 70% of respondents knew researchers who screened their subjects to identify those that responded in a predictable way to various TMS protocols. Fewer respondents admitted to this practice. Among the other questionable practices, 13–21% of respondents had previously failed to report all the experimental conditions from a study, had selectively reported data from sub-sample, of subjects or had rejected data based on a ‘gut’ feeling or without statistical justification. There was nearly total agreement (45 of 47 respondents, 96%) that these sorts of practices should be reported in publications.

Across the 56 papers that were reviewed, none of the questionable research practices noted above were explicitly reported. While the majority of studies monitored background EMG, only four provided clear criteria on the time-period and threshold level used to exclude trials. Only one study included a sample size calculation and several studies did not provide the gender or handedness of subjects.

Several researchers volunteered comments about the survey and their experience using TMS, some of which are highlighted here. One researcher commented on the potential effect of the survey results:

“Thank you for carrying out this study of research practices with TMS. In my opinion it is an area in which real results have become difficult to distinguish from noise due to sloppy research practices (driven by a pressure to publish) and positive publication bias in most journals. It has reached a point where it is very difficult to design any neuromodulation study because the positive control conditions taken from previous results are simply not replicable, almost without exception.”

Another researcher commented on the difficulty to publish negative results and the long-term impact this has on one’s career choice:

“After all my experiences with TMS and troubles publishing negative/smaller results during my PhD, I decided to shift my research career to another subject. I hope something will change in reporting and interpretation of the TMS (but also direct current stimulation) results and the techniques can be of use for some patient groups.”

As a final example, a researcher described their experience with repetitive TMS to improve motor function in patients:

“I perform repetitive TMS for the purpose of functional restoration. In my experience, the traditional 10-session protocol is too weak to induce a robust result. […] I answer this survey based on the observation on traditional 10-session repetitive TMS.”


An anonymous internet-based survey was sent to researchers who had been first or last authors on published papers in which TMS was used to alter motor cortical excitability. The response rate was 31% and many of the respondents had been using TMS for more than a decade. They provided information on research results obtained using a variety of popular TMS techniques and detailed how these studies were designed and carried out. Results from our survey highlight the difficulty experienced by researchers to reproduce published research results. In addition, we found evidence that researchers in this field engage in, but fail to report, questionable research practices.

When questioned about TMS-induced effects on motor cortical excitability, 45–60% of respondents indicated they had success, to a greater or lesser extent, reproducing the original published results. Others were able to reproduce the original results only sometimes (32%) or not at all (10%). While this could be seen as evidence TMS protocols work in certain circumstances, this interpretation may well be wrong. It is well established the neuroscience research is dogged by small sample size, low statistical power and true effects that are often small [10]. Hence the risk of false discoveries is high [18] which causes the size of reported effects to be exaggerated [26, 27] and likely an over-optimistic picture of reproducibility. While some studies and expert reviews have acknowledged the high variability in responses to non-invasive brain stimulation and attempted to isolate controllable factors which determine the variability such as prior motor activity, attention, time of testing, age, and gender [1113, 16, 17], the effect of publication bias, low statistical power and questionable research practices has been largely, if not totally, ignored. Based on our respondents’ reports, up to half of all studies, and hence published papers, in this research field should fail to reproduce the original results. However, such statements are rare in the literature and this biased representation contributes to the false view that TMS-induced effects are robust.

Our survey reveals a varied and somewhat haphazard approach to determining study sample size. About half of respondents relied on personal experience or previously published studies. In about one in five cases, formal sample size calculations were performed, possibly with exaggerated effect sizes from the literature. In a similar fraction of cases, additional subjects were tested based on ad-hoc analyses, or fewer subjects were tested because a clear effect was or was not present. Thus, a sub-set of researchers keep an eye on evolving probability values, a practice that violates a key tenet of a priori statistical testing [31]. Regardless of the approach used, sample sizes are often too small in neuroscience research and this results in low statistical power [10]. All studies can be affected, including those trying to reproduce previously published effects. For example, by testing the same number of subjects as a study that reported barely significant results, you have as much chance of rejecting the same null hypothesis as you do correctly calling a coin flip [10].

But how do studies with small sample sizes discover statistically significant effects? When statistical power is low, the first study to publish an effect is often the most biased towards an extreme result—the winner’s curse [32]. Subsequent studies are often less biased and find evidence of smaller or even contradictory effects [27]. Extreme effects tend to occur when thresholds such as statistical significance are used and they are most severe when studies are too small and thus have low statistical power [10]. For example, the original report on theta-burst stimulation involved only 9 subjects and concluded: ‘Here we describe a very rapid method of conditioning the human motor cortex using rTMS that produces a controllable, consistent, long-lasting, and powerful effect on motor cortex physiology and behavior after an application period of only 20–190 s’ [8]. In stark contrast to these consistent and powerful effects, the same research group conducted a subsequent study involving 52 subjects and concluded: ‘The cTBS and iTBS after-effects were highly variable. Indeed in this set of participants there was no overall effect of either form of stimulation’ [33]. In support of these latter findings, a recent study involving 56 subjects found no overall effect for PAS, iTBS or direct current stimulation [11]. Approximately 40% of subjects responded as expected to each of the stimulation techniques and only 12% of subjects responded in this way to all three techniques. The take-home message is clear: researchers should be wary of new published effects, especially when sample size is small. To establish reproducibility for an effect, a large sample size is mandatory.

There are other, more insidious factors that can bias research results towards extreme values. Questionable research practices is an established term that describes an unhealthy flexibility in collecting and analysing research data. These practices have been documented in many fields (e.g. medicine [23]; psychology [34]) and their incidence is believed to be increasing [21]. Why are such practices so problematic? Computer simulations of experimental data reveal that questionable research practices markedly increase false discovery rates by increasing the likelihood of finding a statistically significant effect (e.g. [24]). In line with previous reports [23, 34], our survey found questionable research practices were not uncommon among researchers. Between 6 and 38% of respondents reported using such practices, whereas 2–3 times as many believed that other researchers in the TMS field used questionable research practices. This mismatch between respondents’ and others’ practices could represent a bias in reporting, or could also be explained by a bias in the sample of respondents who may be especially concerned about questionable research practices, as they elected to complete the survey. The response rate to our survey was similar to that of other web-based surveys [35, 36]. However, it remains to be determined whether our results capture the opinions and experiences of the entire TMS field. The most common questionable practices reported in our survey were to exclude data, select subjects who are known to respond to a TMS technique, choose outcomes and time points, and select data from a sub-group of the main cohort. There was a uniform belief among respondents that such practices should be reported in published reports. In contrast to this sentiment, examination of 56 papers from our original sample found no mention of any questionable research practices. It is unfortunate that these dubious practices which predispose to false positive results appear to be so pervasive that they may constitute everyday practice.

Publication bias—where significant results are more likely to get published—has a long history in science [37]. Approximately half the respondents to our survey indicated they could not reproduce the original TMS effects: is it any wonder we also found evidence of a large file-drawer problem? Many TMS studies are not written up or accepted for publication and end up in a real or metaphorical file-drawer, never to be seen again [20, 27, 38]. This behaviour is well documented across branches of science and has been well analysed in sociology [19, 39], psychology [40, 41] and medicine [42]. While we were surprised by the numbers of dormant papers—an average of 4 per respondent—we were staggered that ∼ 15% of respondents had more papers in their file-drawer than papers published in the TMS literature. This reporting bias needs to be addressed if we are to determine the true effect of TMS techniques. Also worrying are the written comments we received indicating that this reporting bias also impacts the design of clinical interventions and the careers of individual researchers who feel pressured to publish positive results.

A diagram was developed to summarise our findings on the irreproducibility of published TMS results (Fig 1). While not intended to represent all research conducted in this field, the diagram can help identify possible changes to improve the current state of TMS research (indicated in red font). First, researchers have to stop questionable researcher practices. The most severe practices, such as rejecting data based on a gut feeling, have no place in science. The less severe practices are more widespread, with some possibly becoming common practice. Regardless of their perceived severity, questionable research practices tend to not be reported. However, without these details reproducible TMS research will never become a reality [43]. Second, researchers have to stop conducting small studies with questionable sampling practices. The evidence can no longer be ignored and the benefits of larger sample sizes should be welcomed. We should increase the certainty of reported effects [10, 32] rather than doing so cosmetically by using standard error bars on graphs [44]. By increasing the level of certainty surrounding the size of an investigated effect, readers and editors will be interested regardless of the positiveness or negativeness of results, thus doing away with the arbitrary p-value and the file-drawer [32].

Fig 1. Factors contributing to irreproducible TMS results.

When planning and executing a research study, the size of the investigated effect and the size of the sample directly influence a study’s statistical power—probability of correctly rejecting the null hypothesis when the null hypothesis is false—and the certainty of reported researcher results. Selecting sample size based on previous experience, published reports or power calculations based on inflated effect sizes from the literature often results in too few subjects being tested. In a study with low statistical power, significant results (i.e. p < 0.05) are biased towards extreme values(i.e. a large effect; study B). Independently, questionable research practices will also increase the rate of false discoveries and exaggerated effect sizes. Because these results meet the traditional level of statistical significance, they will likely become part of published literature. For the unlucky scientist who did not find statistically significant results (study A), the study may never be written up or it will be rejected by publishers because it presents uncertain, negative results. These studies become part of the cemetery of unpublished scientific research, the file-drawer.

Irreproducible research, publication bias, and questionable research practices have become increasingly worrisome to researchers, policy makers, and journal editors [21, 30, 43, 45, 46]. Our survey reveals that research using TMS to alter motor cortical excitability is not immune to these problems. Commonly, survey respondents were only sometimes able or never able to reproduce the original reported effects, and they and their colleagues engaged in various questionable research practices. While a cure for these large-scale, endemic problems has yet to be found, we believe that increasing study samples size and eliminating—or at least reporting—questionable research practices would be a simple step towards more reproducible TMS research. This is important because TMS research is increasingly being translated into clinical practice, and patients and their physicians have the right to expect that potential therapies are based on sound and reproducible research.

Supporting Information

S1 File. Survey.

Complete survey questions as they appeared to survey respondents.


S3 File. Survey results.

Spreadsheet containing all results from the survey.



We would like to thank the researchers who completed the survey, Dr Andrew Cartwright for developing the online survey infrastructure, and Emily Ainsley for her assistance retrieving e-mail addresses.

Author Contributions

Conceived and designed the experiments: MEH JLT SCG. Performed the experiments: MEH JLT SCG. Analyzed the data: MEH JLT SCG. Contributed reagents/materials/analysis tools: MEH JLT SCG. Wrote the paper: MEH JLT SCG.


  1. 1. Lefaucheur JP, André-Obadia N, Antal A, Ayache SS, Baeken C, Benninger DH, et al. Evidence-based guidelines on the therapeutic use of repetitive transcranial magnetic stimulation (rTMS). Clin Neurophysiol. 2014;125:2150–206. pmid:25034472
  2. 2. Oberman L, Edwards D, Eldaief M, Pascual-Leone A. Safety of theta burst transcranial magnetic stimulation: a systematic review of the literature. J Clin Neurophysiol. 2011;28:67–74. pmid:21221011
  3. 3. Barker AT, Jalinous R, Freeston IL. Non-invasive magnetic stimulation of human motor cortex. Lancet. 1985;1:1106–7. pmid:2860322
  4. 4. Day BL, Rothwell JC, Thompson PD, Maertens de Noordhout A, Nakashima K, Shannon K, et al. Delay in the execution of voluntary movement by electrical or magnetic brain stimulation in intact man. Evidence for the storage of motor programs in the brain. Brain. 1989;112 (Pt 3):649–63. pmid:2731025
  5. 5. Ziemann U, Paulus W, Nitsche MA, Pascual-Leone A, Byblow WD, Berardelli A, et al. Consensus: Motor cortex plasticity protocols. Brain Stimul. 2008;1:164–82. pmid:20633383
  6. 6. Di Lazzaro V, Rothwell JC. Corticospinal activity evoked and modulated by non-invasive stimulation of the intact human motor cortex. J Physiol (Lond). 2014;592:4115–28.
  7. 7. Stefan K, Kunesch E, Cohen LG, Benecke R, Classen J. Induction of plasticity in the human motor cortex by paired associative stimulation. Brain. 2000;123:572–84. pmid:10686179
  8. 8. Huang YZ, Edwards MJ, Rounis E, Bhatia KP, Rothwell JC. Theta burst stimulation of the human motor cortex. Neuron. 2005;45:201–6. pmid:15664172
  9. 9. Martin PG, Gandevia SC, Taylor JL. Theta burst stimulation does not reliably depress all regions of the human motor cortex. Clin Neurophysiol. 2006;117:2684–90. pmid:17029949
  10. 10. Button KS, Ioannidis JPA, Mokrysz C, Nosek BA, Flint J, Robinson ESJ, et al. Power failure: why small sample size undermines the reliability of neuroscience. Nat Rev Neurosci. 2013;14:365–76. pmid:23571845
  11. 11. López-Alonso V, Cheeran B, Río-Rodríguez D, Fernández-del Olmo M. Inter-individual variability in response to non-invasive brain stimulation paradigms. Brain Stimul. 2014;7:372–80. pmid:24630849
  12. 12. Ridding MC, Ziemann U. Determinants of the induction of cortical plasticity by non-invasive brain stimulation in healthy subjects. J Physiol (Lond). 2010;588:2291–304.
  13. 13. Sale MV, Ridding MC, Nordstrom MA. Factors influencing the magnitude and reproducibility of corticomotor excitability changes induced by paired associative stimulation. Exp Brain Res. 2007;181:615–26. pmid:17487476
  14. 14. Nakagawa S, Cuthill IC. Effect size, confidence interval and statistical significance: a practical guide for biologists. Biol Rev Camb Philos Soc. 2007;82:591–605. pmid:17944619
  15. 15. Nieuwenhuis S, Forstmann BU, Wagenmakers EJ. Erroneous analyses of interactions in neuroscience: a problem of significance. Nat Neurosci. 2011;14:1105–7. pmid:21878926
  16. 16. Vallence AM, Ridding MC. Non-invasive induction of plasticity in the human cortex: uses and limitations. Cortex. 2014;58:261–71. pmid:24439754
  17. 17. Vallence AM, Goldsworthy MR, Hodyl NA, Semmler JG, Pitcher JB, Ridding MC. Inter- and intra-subject variability of motor cortex plasticity following continuous theta-burst stimulation. Neuroscience. 2015 Jul 21 pii: S0306-4522(15)00657-0. pmid:26208843
  18. 18. Ioannidis JPA. Why most published research findings are false. PLoS Med. 2005;2:e124. pmid:16060722
  19. 19. Mervis J. Research Transparency. Why null results rarely see the light of day. Science. 2014;345:992. pmid:25170131
  20. 20. Rosenthal R. The file drawer problem and tolerance for null results. Psychol bull. 1979;86:638–41.
  21. 21. Anderson WP. Reproducibility: Stamp out shabby research conduct. Nature. 2015;519:158. pmid:25762270
  22. 22. Dwan K, Gamble C, Williamson PR, Kirkham JJ. Systematic review of the empirical evidence of study publication bias and outcome reporting bias—an updated review. PLoS ONE. 2013;8:e66844. pmid:23861749
  23. 23. Martinson BC, Anderson MS, de Vries R. Scientists behaving badly. Nature. 2005;435:737–8. pmid:15944677
  24. 24. Simmons JP, Nelson LD, Simonsohn U. False-positive psychology: undisclosed flexibility in data collection and analysis allows presenting anything as significant. Psychol Sci. 2011;22:1359–66. pmid:22006061
  25. 25. Head ML, Holman L, Lanfear R, Kahn AT, Jennions MD. The extent and consequences of p-hacking in science. PLoS Biol. 2015;13(3):e1002106. pmid:25768323
  26. 26. Ioannidis JPA. Why most discovered true associations are inflated. Epidemiology. 2008;19:640–8. pmid:18633328
  27. 27. Schooler J. Unpublished results hide the decline effect. Nature. 2011;470:437. pmid:21350443
  28. 28. Kaplan RM, Irvin VL. Likelihood of Null Effects of Large NHLBI Clinical Trials Has Increased over Time. PLoS ONE. 2015;10(8):e0132382. pmid:26244868
  29. 29. Holman L, Head ML, Lanfear R, Jennions MD. Evidence of Experimental Bias in the Life Sciences: Why We Need Blind Data Recording. PLoS Biol. 2015;13(7):e1002190. pmid:26154287
  30. 30. Alberts B, Cicerone RJ, Fienberg SE, Kamb A, McNutt M, Nerem RM, et al. Self-correction in science at work. Science. 2015;348(6242):1420–2. pmid:26113701
  31. 31. Kline RB. Beyond significance testing: Statistical reform in the behavioral sciences. 2nd ed. Washington, DC: American Psychological Association; 2013.
  32. 32. Halsey LG, Curran-Everett D, Vowler SL, Drummond GB. The fickle P value generates irreproducible results. Nat Methods. 2015;12(3):179–85. pmid:25719825
  33. 33. Hamada M, Murase N, Hasan A, Balaratnam M, Rothwell JC. The role of interneuron networks in driving human motor cortical plasticity. Cereb Cortex. 2013;23:1593–605. pmid:22661405
  34. 34. John LK, Loewenstein G, Prelec D. Measuring the prevalence of questionable research practices with incentives for truth telling. Psychol Sci. 2012;23:524–32. pmid:22508865
  35. 35. Hayslett MM, Wildemuth BM. Pixels or pencils? The relative effectiveness of Web-based versus paper surveys. Library & Information Science Research. 2004;26:73–93.
  36. 36. Cunningham CT, Quan H, Hemmelgarn B, Noseworthy T, Beck CA, Dixon E, et al. Exploring physician specialist response rates to web-based surveys. BMC Med Res Methodol. 2015;15–32:32.
  37. 37. Sterling TD. Publication Decisions and Their Possible Effects on Inferences Drawn from Tests of Significance–Or Vice Versa. J Am Stat Assoc. 1959;54(285):30–5.
  38. 38. Simonsohn U, Nelson LD, Simmons JP. P-curve: a key to the file-drawer. J Exp Psychol Gen. 2014;143:534–47. pmid:23855496
  39. 39. Franco A, Malhotra N, Simonovits G. Publication bias in the social sciences: unlocking the file drawer. Science. 2014;345:1502–5. pmid:25170047
  40. 40. Ferguson CJ, Brannick MT. Publication bias in psychological science: prevalence, methods for identifying and controlling, and implications for the use of meta-analyses. Psychol Methods. 2012;17:120–8. pmid:21787082
  41. 41. Bakker M, van Dijk A, Wicherts JM. The Rules of the Game Called Psychological Science. Perspect Psychol Sci. 2012;7(6):543–54. pmid:26168111
  42. 42. McGauran N, Wieseler B, Kreis J, Schüler YB, Kölsch H, Kaiser T. Reporting bias in medical research—a narrative review. Trials. 2010;11:37. pmid:20388211
  43. 43. Nosek BA, Alter G, Banks GC, Borsboom D, Bowman SD, Breckler SJ, et al. Promoting an open research culture. Science. 2015;348(6242):1422–5. pmid:26113702
  44. 44. Cumming G. The new statistics: Why and how. Psychol Sci. 2014;25:7–29. pmid:24220629
  45. 45. Begley GC, Buchan AD, Dirnagl U. Institutions must do their part for reproducibility. Nature. 2015;525:25–7. pmid:26333454
  46. 46. Nuzzo R. Scientific method: statistical errors. Nature. 2014;506(7487):150–2. pmid:24522584