Peer Review History
| Original SubmissionOctober 6, 2025 |
|---|
|
PONE-D-25-53826Outliers (typically) cannot cause type I errors in one-sample / paired t-testsPLOS ONE Dear Dr. Wisler, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Dec 18 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Abhik Ghosh Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. If the reviewer comments include a recommendation to cite specific previously published works, please review and evaluate these publications to determine whether they are relevant and should be cited. There is no requirement to cite these works unless the editor has indicated otherwise. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Partly Reviewer #3: Yes Reviewer #4: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes Reviewer #4: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes Reviewer #4: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes Reviewer #4: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: This manuscript tackles a pervasive and often misunderstood problem in statistical inference with a commendable blend of theoretical derivation and simulation. The findings are potentially significant for guiding empirical research. However, the work requires major revision to enhance its clarity, robustness, and practical applicability. The assumptions are quite restrictive, and the presentation of the core theorems could be made more accessible to the broader target audience of applied researchers. 1. The assumption of unit variance (σ^x2=1σ^x2=1), while mathematically convenient for derivation, limits the immediate interpretability for applied researchers. The manuscript should include a more detailed discussion or a supplementary section on how to apply these bounds to real-world data where this assumption does not hold, perhaps with a worked example. 2. The definition of an outlier "causing" rejection is logically sound but operationally challenging. The manuscript should discuss the practical implications of this definition, especially in scenarios where multiple outliers might be present, and how a practitioner might distinguish between an outlier-caused discovery and a genuine one. 3. The simulation is well-designed but could be strengthened by testing the robustness of the derived bounds against violations of normality. Including data generated from heavy-tailed distributions (e.g., t-distribution) would demonstrate the generalizability of the findings beyond the ideal Gaussian case assumed in the model. 4. The connection between the theoretical bounds and practical outlier detection rules (e.g., 2-σ or 3-σ) is a key contribution. However, the discussion would be enriched by engaging with more modern, robust statistical methods. The authors are strongly encouraged to review Robust Statistical Methods for Handling Outliers in High-Dimensional Data, Journal of the American Statistical Association; A robust identification method for stochastic nonlinear parameter varying systems, Mathematical Modelling and Control; Extended state observer based fractional order sliding mode control for steer-by-wire systems, IET Control Theory and Applications. These works demonstrate advanced techniques for handling uncertainties and outliers in dynamic systems, which could provide a broader context for the statistical problem addressed here and inspire more robust interpretations. 5. The manuscript concludes that the risk is "negligible" in many situations. This is a strong claim that should be tempered. A more nuanced conclusion, acknowledging the specific and narrow conditions under which outliers can cause type I errors, would be more accurate and helpful for the reader. 6. The abstract and introduction effectively set up the problem, but the "Methods and Results" section is dense and could benefit from additional explanatory text to guide the reader through the logical flow of the derivations, making the paper more accessible to those less versed in mathematical statistics. Reviewer #2: The manuscript titled “Outliers (typically) cannot cause Type I errors in one-sample/paired t-tests” presents a technically rigorous and mathematically sound analysis of how outliers influence Type I error rates in t-tests. The work is well motivated and addresses an important and long-debated topic in applied statistics. The author combines theoretical derivations with simulation evidence to demonstrate that, under controlled assumptions, outliers rarely inflate Type I errors. The findings are internally consistent and supported by the presented data. That said, the scope of the contribution is somewhat limited, and the paper would benefit from several substantive and editorial improvements before it can meet the standards of PLOS ONE. My detailed observations and suggestions are as follows: 1. Technical Strengths The mathematical framework and derivations are accurate, logically consistent, and presented in a reproducible manner. The Monte Carlo simulations are appropriate in design and provide convincing numerical support for the theoretical results. The results contribute to the broader discussion of robustness in classical parametric testing, clarifying misconceptions about the role of outliers in Type I error inflation. 2. Limitations and Areas for Improvement a. Scope and Novelty: While the results are interesting, the contribution is incremental rather than groundbreaking. Several prior studies (e.g., Zimmerman, 1994; Derrick et al., 2017; Wilcox, 2020) have reached similar conclusions through simulation-based approaches. The current work’s novelty lies mainly in providing theoretical bounds (Δ₁, Δ₂), which could be valuable but do not represent a major conceptual advance for a multidisciplinary audience. b. Generalizability of Findings: The analysis is confined to data simulated from normal distributions with fixed variance and controlled outlier contamination. To strengthen generalizability, consider: Including non-normal or heavy-tailed distributions (e.g., t-distribution, log-normal) to demonstrate robustness. Discussing whether similar results hold under heteroscedasticity or when data violate independence assumptions. c. Empirical Relevance: All results are based solely on simulations. Including a short applied example—even using an open dataset from behavioral or biomedical research—would substantially enhance the practical relevance and reader engagement. This could also highlight real-world implications of the derived bounds. d. Presentation and Clarity: The manuscript is generally well written but would benefit from professional language editing to correct typographical errors, long sentences, and redundant phrasing. Figures are informative, but some could use clearer legends and axis labels. Ensure consistent notation for mathematical symbols and parameters throughout (e.g., Δ₁, Δ₂, μ̂, σ̂). The abstract should explicitly summarize the methodological contribution and practical implications rather than focusing solely on background. e. Literature Review: The literature review is appropriate but can be enhanced by citing and contrasting recent robust statistical inference studies (2022–2024), including methods involving M-estimators, bootstrap-based robust t-tests, or influence function analyses. This would contextualize the current findings within the evolving field of robust statistics. 3. Data Availability and Reproducibility The author appropriately provides simulation code and data availability details via an OSF repository, aligning with PLOS ONE’s open-data policy. However, please ensure the exact repository link is included in the Data Availability Statement and that sufficient details (software version, random seeds, replication size) are specified for full reproducibility. 4. Overall Evaluation The study is methodologically solid, mathematically correct, and clearly motivated. However, the breadth of contribution and applied impact are limited, making the work more appropriate for a specialized statistical methods journal than a general multidisciplinary platform like PLOS ONE. If the author expands the simulation scope, improves clarity, and adds an applied example, the manuscript could become suitable for reconsideration. Summary Recommendation Based on the current form, I recommend rejection due to limited novelty and restricted scope for the broad readership of PLOS ONE. However, the work is of good technical quality and may merit publication in a specialized statistical journal after revision and expansion. Reviewer #3: Overall Assessment The paper “Outliers (typically) cannot cause type I errors in one-sample/paired t-tests” is technically sound and well written. The introduced mathematical bounds are both simple and ingenious, and the results are presented clearly. However, several aspects of the manuscript could be improved to enhance clarity, completeness, and presentation. Major Concerns 1. Introduction – Insufficient bibliography The introduction would benefit from a more comprehensive review of related work. For instance, in line 49, additional examples or references could be added. 2. Introduction – Missing paper outline The manuscript lacks a clear outline of its structure. A brief description of the organization (e.g., “Section 2 introduces…, Section 3 presents…”) would guide readers and improve readability. 3. Lack of empirical validation The article would benefit from including an analysis of a real dataset where the proposed bounds can be observed in practice. Minor Concerns 4. Figure 1 – Axis readability The axes in Figure 1 are too small. 5. Discussion – Use of subjective adjectives Avoid using subjective or qualitative adjectives such as “small-to-moderate,” “large,” “not too large,” “good,” and “bad” when discussing the results. More precise or quantitative descriptions are recommended. Reviewer #4: In this paper, the authors investigate whether a single outlier can cause the rejection of the null hypothesis in a one-sample t-test. The main contributions are two theorems that characterize the magnitude of potentially harmful outliers. First, the authors show that outliers above a certain magnitude can only decrease the test statistic and therefore cannot cause type I errors. Second, they derive an explicit expression for the outlier magnitude that maximizes the effect on the test statistic. The results are further discussed and illustrated with Monte Carlo simulations. The topic of outlier handling is timely and important. To the best of my knowledge, the results presented in this paper are correct, novel, and relevant both in theory and in practice. The paper is well written, the proofs are transparent and the simulation results convincing. Overall, I found this to be a very interesting and enjoyable read, and I recommend the paper for publication. I have, however, a few comments and suggestions. Some of them may go beyond the intended scope of the paper, so I leave it to the author’s discretion which ones to incorporate. 1. The approach taken in the paper seems closely related to ideas from robust statistics, in particular to (empirical) influence functions. The problem considered here could likely be embedded in that framework by viewing the t-statistic as an estimator of a standardized mean difference and then studying its sensitivity to point-mass contamination. Classic treatments of this perspective can be found in [1]. Referring to this line of work would help situate the problem and show that the present result is consistent with a more general robustness perspective. That said, I am not sure how far it is worth pursuing this in the current paper — in the worst case, it may overcomplicate an otherwise clean argument. Still, I would recommend at least pointing out that this connection exists. 2. Some results are stated somewhat implicitly. For example, I understand that Δ_1^* > Δ_2^*. Is this correct If so, it would be helpful to state and prove it explicitly. Also, is the inequality strict for some (or all) sample sizes and sample means? Based on Theorem 2, it should also be possible to give an explicit expression for the maximum difference t' - t. 3. It might help some readers to add a plot of t' (and the original t, as a constant) as functions of Δ for fixed μ^ and n. Highlighting Δ_1^* and Δ_2^* in this plot would make the results more transparent. 4. Intuitively, one might expect smaller sample sizes to lead to tests that are more sensitive to outliers. Here, the paper shows a somewhat counterintuitive phenomenon: extreme outliers can cause type I errors only for sufficiently large sample sizes. At the same time, the area in the μ^-Δ plane in which outliers cause type I errors does seem to decrease as the sample size increases. Maybe the author could expand the discussion on these effects. 5. A practically very relevant question is: Given a significance level and an observed t-value that leads to a rejection of the null hypothesis, is it possible that this rejection was caused by an outlier in the data? My impression is that the results in this paper can be used to answer this question. 6. As the author notes, extending the analysis to multiple outliers is a natural next step. The analogue of comment 5 would then be: How many data points would need to be outliers for the rejection of the null hypothesis to be possibly caused to them? 7. Minor issues: - In the derivation of the “New Sample Variance,” the cross term in the third line is incorrect. The final result is still correct because the term is zero, but the intermediate line should be fixed. - In line 156, a word seems to be missing: “the data cannot [increase?] the t-statistic.” [1] Hampel, F. R., Ronchetti, E. M., Rousseeuw, P. J., & Stahel, W. A. (1986), Robust Statistics: The Approach Based on Influence Functions (Wiley) ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No Reviewer #3: No Reviewer #4: No ********** [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] To ensure your figures meet our technical requirements, please review our figure guidelines: https://journals.plos.org/plosone/s/figures You may also use PLOS’s free figure tool, NAAS, to help you prepare publication quality figures: https://journals.plos.org/plosone/s/figures#loc-tools-for-figure-preparation. NAAS will assess whether your figures meet our technical requirements by comparing each figure against our figure specifications. |
| Revision 1 |
|
Outliers (typically) cannot cause type I errors in one-sample / paired t-tests PONE-D-25-53826R1 Dear Dr. Wisler, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. For questions related to billing, please contact billing support. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Abhik Ghosh Academic Editor PLOS One Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed Reviewer #2: All comments have been addressed Reviewer #3: All comments have been addressed Reviewer #4: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes Reviewer #4: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes Reviewer #4: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes Reviewer #4: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes Reviewer #4: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: (No Response) Reviewer #2: Strengths 1. Strong theoretical contribution with clear derivations and formal proofs. 2. Well-designed simulations that directly validate theoretical results. 3. High relevance to applied researchers who routinely face decisions about outlier handling. 4. Excellent reproducibility, with simulation code openly available. ________________________________________ Minor Concerns / Suggestions 1. Title tone The word “typically” in the title is appropriate but may be perceived as informal or ambiguous. The author may consider briefly clarifying its meaning in the introduction. 2. Practical guidance While the theoretical insights are strong, the manuscript could benefit from a short subsection translating results into practical recommendations for applied researchers (e.g., when outlier removal is unlikely to inflate Type I error). 3. Discussion length The Discussion section is thorough but could be slightly condensed to improve readability without loss of content. 4. Terminology consistency The interchangeable use of “one-sample” and “paired” t-tests is justified but may briefly confuse non-technical readers; a reminder statement may help. These are minor issues and do not affect the scientific validity of the work. Reviewer #3: The comments for the paper titled "Outliers (typically) cannot cause type I errors in one-sample / paired t-tests" were properly addresed. Reviewer #4: I appreciate the revised manuscript and the detailed responses to the reviewers’ comments. Overall, the paper has been significantly improved. The presentation is clearer, the generality of the results is made more explicit, the work is better positioned in the wider context of robust statistics, and the expanded bibliography is much more comprehensive. Three minor comments: - Figure 1 is a useful addition to the paper and provides a lot of information. It may in fact provide too much information, to the extend that it appears visually overloaded. I would suggest simplifying and decluttering the figure. For example, the number of curves could be reduced, the visualization of \Delta_1 and \Delta_2 could be simplified, and the derivative shown in the rightmost panel seems unnecessary for the discussion of the results. I believe that a simpler plot, such as the one in the attached file, would be sufficient to illustrate \Delta_1 and \Delta_2. The dependence of the latter on \mu^ and n could be shown in a separate figure, if at all. - On rereading the paper, I agree with the comments that the language could be more objective and precise in places, even in the revised version. For example, in the abstract, it is unclear whether the term "minimal" is used informally (to mean small or negligible) or in the formal sense of "attaining a minimum". Similarly, it is not clear what is meant by describing the experiments as "unique." More generally, words such as "actually" are rarely necessary in formal writing. One concrete suggestion: the phrase "we impose the following two conditions, that do not actually restrict the generality of the analysis" could be replaced with "without loss of generality, we assume that..." - I found the extended discussion section very helpful. One point addressed there that could be highlighted earlier is that the paper focuses specifically on whether outliers (extreme values) in the data can lead to false positives. It does not address the more general question of the sensitivity of the t-test to other forms of model mismatch or data contamination. Clarifying this distinction earlier in the paper may help avoid confusion. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No Reviewer #3: No Reviewer #4: No **********
|
| Formally Accepted |
|
PONE-D-25-53826R1 PLOS One Dear Dr. Wisler, I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS One. Congratulations! Your manuscript is now being handed over to our production team. At this stage, our production department will prepare your paper for publication. This includes ensuring the following: * All references, tables, and figures are properly cited * All relevant supporting information is included in the manuscript submission, * There are no issues that prevent the paper from being properly typeset You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps. Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. You will receive an invoice from PLOS for your publication fee after your manuscript has reached the completed accept phase. If you receive an email requesting payment before acceptance or for any other service, this may be a phishing scheme. Learn how to identify phishing emails and protect your accounts at https://explore.plos.org/phishing. If we can help with anything else, please email us at customercare@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Abhik Ghosh Academic Editor PLOS One |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .