Peer Review History
| Original SubmissionNovember 7, 2024 |
|---|
|
Dear Dr. Habibzadeh, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by May 03 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols . We look forward to receiving your revised manuscript. Kind regards, Christine E. King, PhD Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, we expect all author-generated code to be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse. 3. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? Reviewer #1: Partly Reviewer #2: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? -->?> Reviewer #1: Yes Reviewer #2: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available??> The PLOS Data policy Reviewer #1: Yes Reviewer #2: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English??> Reviewer #1: Yes Reviewer #2: Yes ********** Reviewer #1: The computer simulation presented here is based on an interesting and sensible premise: It might be sensible to flexibly adjust the alpha level (or, “PST”, in the author’s terminology) employed in statistical significance testing in order to optimize the risks of type-1 and type-2 errors. I find most of the manuscript clear and the basic premise of the simulations convincing. However, I think that a number of issues need to be addressed. 1) I find the title unhelpful. There are many ways to determine if a replication was successful. This does not need to rely on statistical significance. I would find title that focusses on flexible adjustment of PST much more helpful. 2) The introduction is somewhat myopic. Previous discussions of the topic by other authors should be acknowledged (e.g., https://daniellakens.blogspot.com/2019/05/justifying-your-alpha-by-minimizing-or.html and references therein). 3) It would be helpful to mention early on that this paper focusses on dvs. Similarly, the important concept of weighting type-1 vs. type-2 errors (i.e., one type of error might be considered more harmful than the other) should be appropriately introduced in the introduction. An example might be helpful. 4) The introduction of unequal variances in the two simulated study arms puzzled me. What is the motivation behind it? Is this meant as a device to manipulate the true effect size (ES)? Manipulation of the ES via the difference in the population means would be much easier to understand (i.e., the difference in population means equals d when the true variance for both populations remains fixed at 1). Also, with unequal variances researchers might not rely on the t-test for analysis as is assumed in the simulation. 5) I find the data space explored by the simulation overly narrow. It would be good to systematically manipulate not just N and ES but also the relative weight given to type-1 and type-2 errors and the base rate of these errors. Dealing with results from a design with that many factors (including their interactions, which are currently neglected) is challenging, but you might find this approach helpful https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0262809. 6) I don’t understand how the misclassification rate can exceed 52.5% for alpha = .05 (see Figure 1). Simulations with H0 being true lead to 2.5% misclassifications (alpha x 50% H0 base rate). In the most extreme case that statistical power = zero, simulations with H1 being true add another 50% misclassifications (beta x 50% H1 base rate). Both add to just 52.5%. I therefore wonder if an error crept into the simulation. 7) I did not understand Figure 3 and the related arguments in the discussion. Without further arguments, I don’t find it problematic when determination of the optimal PST relies on observed sample statistics. 8) In a surprising swerve, the discussion argues of the superiority of Bayes over frequentist statistics. I find that unhelpful because the simulations allow no such comparison. Johannes Hönekopp Reviewer #2: Dear Authors, Thank you for submitting your manuscript, On the Reproducibility of Randomized Clinical Trials, to PLOS ONE. Your work addresses a critical issue in modern scientific research and provides valuable insights into the limitations of frequentist statistical frameworks. Below, I outline my evaluation and suggestions for improvement to strengthen the manuscript further. Your Monte Carlo simulation study is well-designed and rigorously executed, offering a compelling comparison of fixed vs. flexible p-value thresholds. The conclusions are largely supported by the data. However, the discussion of Bayesian methods as an alternative approach remains underdeveloped. Please provide concrete examples or references illustrating how Bayesian methods (e.g., posterior probability calculations, Bayes factors) might resolve the conflicts inherent in frequentist frameworks, and discuss how Bayesian approaches address reproducibility challenges in practical terms (e.g., prior incorporation, decision thresholds). The statistical methodology is robust, please clarify why specific parameters (e.g., prior probability pr=0.5, effect size = 0.5) were chosen. A sensitivity analysis exploring how varying these parameters impacts the flexible threshold’s performance would strengthen the study’s applicability. Finally, please highlight practical challenges of implementing flexible thresholds (e.g., a posteriori calculation, variability across replicas) in the discussion section. This study makes a meaningful contribution to the discourse on reproducibility and statistical inference. Thank you for the opportunity to review your work. Best regards, ********** what does this mean? ). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy Reviewer #1: Yes: Johannes Hönekopp Reviewer #2: Yes: Angel A. García O'Diana ********** [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org . Please note that Supporting Information files do not need this step. |
| Revision 1 |
|
Dear Dr. Habibzadeh, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. 1) The use of fixed prior probabilities (pr = 0.5) and seriousness ratios (C = 0.25) needs to be tested through sensitivity analysis. Robustness under varied assumptions would strengthen the manuscript’s generalizability. 2) The assumption of pooled variance in t-testing under heteroscedasticity requires stronger justification or a robustness check via Welch’s t-test. 3) The manuscript should comment on convergence diagnostics for the optimization routine (optim()), since flexible PST values are highly sensitive to this step. Please submit your revised manuscript by Jun 20 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.
If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols . We look forward to receiving your revised manuscript. Kind regards, Christine E. King, PhD Academic Editor PLOS ONE Journal Requirements: Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author Reviewer #1: All comments have been addressed Reviewer #2: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions??> Reviewer #1: Yes Reviewer #2: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? -->?> Reviewer #1: Yes Reviewer #2: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available??> The PLOS Data policy Reviewer #1: Yes Reviewer #2: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English??> Reviewer #1: Yes Reviewer #2: Yes ********** Reviewer #1: (No Response) Reviewer #2: Dear Dr. Habibzadeh, Thank you for your thoughtful and well-executed manuscript on the effect of flexible adjustment of the p-value significance threshold on the reproducibility of randomized clinical trials. Your approach offers a valuable contribution to the ongoing discussion surrounding statistical inference and reproducibility, particularly by proposing a flexible significance threshold based on minimizing a weighted sum of Type I and Type II errors. While your manuscript is methodologically robust and generally well-presented, I would like to offer the following comments and suggestions to strengthen the manuscript further prior to publication: Sensitivity Analysis of Prior Probability and Error Seriousness Coefficient The current simulations assume a fixed prior probability (pr = 0.5) and a seriousness ratio (C = 0.25). While these values may reflect common scenarios in clinical research, their generalizability across fields is limited. ➤ I recommend adding a sensitivity analysis varying both pr (e.g., 0.2, 0.8) and C (e.g., 0.1, 1.0). This would enhance the robustness of your conclusions and demonstrate the flexibility and limitations of the proposed method under different decision-theoretic contexts. Use of Pooled Variance under Heteroscedasticity The simulations employ Student’s t-test with pooled variance, even in scenarios where variances between groups are unequal (s2/s1 ≠ 1). ➤ Please justify this methodological choice more clearly. Alternatively, consider including results using Welch’s t-test as a robustness check. This would provide greater assurance that conclusions drawn under the assumption of pooled variance are not overly sensitive to violations of homoscedasticity. Convergence Validation of the Optimization Routine (optim) Your method relies heavily on the optim() function to estimate flexible thresholds by minimizing a cost function. While this is an appropriate approach, the manuscript does not discuss whether convergence was reliably achieved or how non-convergence was handled. ➤ I suggest including either a brief description of the convergence diagnostics performed (e.g., convergence rate, error handling), or augmenting the code with a validation step to report non-converged iterations. These refinements will bolster the scientific rigor and transparency of your manuscript, improving its reproducibility and interpretability for a broad audience. Kind regards, ********** what does this mean? ). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy Reviewer #1: No Reviewer #2: Yes: Angel Alfonso García O'Diana ********** [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org |
| Revision 2 |
|
On the effect of flexible adjustment of the p value significance threshold on the reproducibility of randomized clinical trials PONE-D-24-50574R2 Dear Dr. Habibzadeh, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Christine E. King, PhD Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: |
| Formally Accepted |
|
PONE-D-24-50574R2 PLOS ONE Dear Dr. Habibzadeh, I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team. At this stage, our production department will prepare your paper for publication. This includes ensuring the following: * All references, tables, and figures are properly cited * All relevant supporting information is included in the manuscript submission, * There are no issues that prevent the paper from being properly typeset You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps. Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. If we can help with anything else, please email us at customercare@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Christine E. King Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .