Peer Review History
Original SubmissionAugust 19, 2020 |
---|
PONE-D-20-25993 A computational reproducibility study of PLOS ONE articles featuring longitudinal data analyses PLOS ONE Dear Dr. Seibold, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. The two reviewers and I agree that your study was well done, clearly reported, and relevant, and that your work contributes to our understanding of reproducibility of studies using longitudinal analyses. Yet, several issues need to be dealt with in your revision. First, the two reviewers requested some additional clarity of the type of submission (registered report or not). I apologize for not indicating to the reviewers that your work was originally submitted as a registered report but later handled as a standard submission. Please clarify this. Second, both reviewers asked you to provide additional details of the methods, sampling, operationalizations, and data access. Both reviewers provided detailed feedback on the reporting and analyses that I ask you to consider as you revise your submission. Third, the reviewers indicate that the main goals and main results could be presented more clearly in several sections of the manuscript. It is also import to be clear on your definition of reproducibility as you present the main results in the abstract. Please submit your revised manuscript by Nov 16 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols We look forward to receiving your revised manuscript. Kind regards, Jelte M. Wicherts Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. We suggest you thoroughly copyedit your manuscript for language usage, spelling, and grammar. If you do not know anyone who can help you do this, you may wish to consider employing a professional scientific editing service. Whilst you may use any professional scientific editing service of your choice, PLOS has partnered with both American Journal Experts (AJE) and Editage to provide discounted services to PLOS authors. Both organizations have experience helping authors meet PLOS guidelines and can provide language editing, translation, manuscript formatting, and figure formatting to ensure your manuscript meets our submission guidelines. To take advantage of our partnership with AJE, visit the AJE website (http://learn.aje.com/plos/) for a 15% discount off AJE services. To take advantage of our partnership with Editage, visit the Editage website (www.editage.com) and enter referral code PLOSEDIT for a 15% discount off Editage services. If the PLOS editorial team finds any language issues in text that either AJE or Editage has edited, the service provider will re-edit the text for free. Upon resubmission, please provide the following:
3. In your manuscript text, we note that "We did not choose the papers randomly, but based on the set of potential papers given to us by PLOS ONE and then selected all papers meeting our criteria". In the Methods, please ensure that you have included the following: - Information about how the initial set of 57 papers was selected, including any inclusion/exclusion criteria applied, so that other interested researchers can reproduce this analysis. We would also recommend that the complete set of 57 initial papers be provided as a supplementary information file. - The complete inclusion/exclusion criteria used to select the initial set of 14 papers from the 57 papers that were identified. - The complete inclusion/exclusion criteria used to select the final set of 11 papers. 4. Please amend the manuscript submission data (via Edit Submission) to include authors Severin Czerny, Siona Decke, Roman Dieterle, Thomas Eder, Steffen Fohr, Nico Hahn, Rabea Hartmann, Christoph Heindl, Philipp Kopper, Dario Lepke, Verena Loidl, Maximilian Mandl, Sarah Musiol, Jessica Peter, Alexander Piehler, Elio Rojas, Stefanie Schmid, Hannah Schmidt, Melissa Schmoll, Lennart Schneider, Xiao-Yin To, Viet Tran, Antje Volker, Moritz Wagner, Joshua Wagner, Maria Waize, Hannah Wecker, Rui Yang, Simone Zellner. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: I Don't Know ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: No Reviewer #2: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Manuscript ID: PONE-D-20-25993 Manuscript title: A computational reproducibility study of PLOS ONE articles featuring longitudinal data analyses Summary This article reports a retrospective observational study designed to test the analytic reproducibility of a small set of PLOS ONE articles containing longitudinal analyses. Specifically, as a part of a class exercise, the authors and their students attempted to repeat the original analyses performed in the selected articles to see if they could obtain similar results. A range of difficulties reproducing the original results were encountered – some of which could be resolved through contact with the original authors. The generalizability of the results is quite limited due to the small sample size and somewhat ad-hoc sampling procedures; however, the authors appropriately calibrate their conclusions to the evidence, for example stating that “We can and should not draw conclusions from our findings on the 11 selected papers on the broader scientific landscape.” Generally, the paper is clearly written and is concise, though I think some important information is absent (see detailed comments below). The study appears to be transparently reported with analysis scripts and data for each reproducibility attempt made publicly available in an Open Science Framework repository. I only checked this repository superficially – one issue is that I could not seem to identify a data file for the study (see comment below) which needs to be addressed. Important note: After reading the paper and writing my review, I was surprised to find a document included after the reference section with the title “Review and response to the registered report”. I was not informed that this study was a registered report and as far as I can tell this is not mentioned in the manuscript aside from this appended document. Can the circumstances of this study be clarified? If this is a registered report then I feel I should be given more information. Most importantly, I need to see the original registered protocol. I would also like to see the prior review history and know whether the original stage 1 reviewers are also appraising the stage 2 manuscript. Major Comments - The exact operationalization of several of the outcome variables could be made clearer. For example, for the question “Are the methods used clearly documented in paper or supplementary material (e.g. analysis code)?” what was considered ‘clear’ vs. ‘unclear’? For “Do we receive the same (or very similar) numbers in tables, figures and models?” what was considered ‘similar’ vs. ‘dissimilar’? For “What are characteristics of papers which make reproducibility easy/possible or difficult/impossible?” – exactly what characteristics were examined? - The oversight of the student work could be described in more detail – did teachers fully - In the methods section, the sampling procedure is somewhat unclear – for example, “57 papers were initially screened. The PLOS website search function was utilized to scan through PLOS ONE published works. Key words used were “mixed model”, “generalized estimating equations”, “longitudinal study” and “cohort study”. 14 papers fulfilled the criteria and were selected”. Where does the number 57 come from? For the 14 papers – did they fulfill the search criteria? Or the criteria in Fig 1? Or both? - It seems to me that the selection criteria will have heavily biased the results in favour of positive reproducibility outcomes – for example, only studies that had confirmed data availability were selected and only studies where authors replied to contact and were favourable to the reproducibility attempt were included. Because these factors probably influenced the results quite substantially, I’d suggest this bias is mentioned in key parts of the paper like the abstract and introduction. - I examined the OSF repository (https://osf.io/xqknz/) for this study and it was unclear to me where to find the data for this study (I could find data for the studies that the authors attempted to reproduce). Could clear instructions be provided on how to find and access the study data? - Could it be clarified if all analyses reported in eligible papers were examined or just the subset pertaining to longitudinal analyses? - It is reported that for one article partial analysis code was available and for a second article the full analysis code was made available during email exchanges with the original authors. Its not clear whether all original authors were explicitly asked to make their code available – could this be clarified? If so, what were the responses to this query? Were reasonable justifications provided for not sharing the original code? - The operational definition of reproducibility could be made clearer in the methods section (and perhaps also in the introduction) – in the results section the authors state “we define similar results as having the same implied interpretations” – this seems to be a less strict definition than used in other studies of analytic reproducibility (e.g., Hardwicke et al., 2018; 2020; Minocher et al., 2020). Some clarification and comment on this would be helpful for understanding the results. - I think it would be informative to mention in the abstract how many analyses were reproducible only when assistance was provided by original authors. - This sentence in the discussion is unclear – “We did not choose the papers randomly, but based on the set of potential papers given to us by PLOS ONE and then selected all papers meeting our criteria (see Figure 1).” If the papers were given to the authors by PLOS ONE then this needs to be mentioned and explained at least in the methods section. Minor Comments - Terminology usage in this domain is diverse and sometimes contradictory (see e.g., https://doi.org/10.1126/scitranslmed.aaf5027) – I’d recommend including explicit definitions and avoiding use of loaded terminology if possible. For example, it would be good to have a clear definition of ‘computational reproducibility’ in the opening paragraph. The authors may also want to consider using the term ‘analytic reproducibility’ instead of computational reproducibility. Researchers in this domain have recently started to draw a distinction between the two concepts and the former seems more applicable to what the present study has addressed. The distinction is discussed in this article - https://doi.org/10.31222/osf.io/h35wt – specifically, “Computational reproducibility is often assessed by attempting to re-run original computational code and can therefore fail if original code is unavailable or non-functioning (e.g., Stodden et al., 2018; Obels et al., 2019). By contrast, analytic reproducibility is assessed by attempting to repeat the original analysis procedures, which can involve implementing those procedures in new code if necessary (e.g., Hardwicke et al., 2018; Minocher et al., 2020)” - An additional point on terminology - use of the term ‘replication’ (e.g., in the abstract and introduction) should perhaps be avoided if possible in this context because it is often used to mean “repeating original study methods and obtaining new data” – whereas here it is being used synonymously with computational reproducibility to mean “repeating original study analyses on the original data” (see http://arxiv.org/abs/1802.03311) - I felt the study design could be made much more explicit in the introduction. For example, “The articles we chose are [1–11]” – briefly mentioning the sampling procedure would be helpful here so the reader can understand the study design (e.g., was it a random sample, arbitrary sample, etc). - The rationale for the study could be made clearer in the introduction. The review of existing literature in this domain is sparse – it is not clear what knowledge gap the study is trying to fill. How does this work build on previous studies and/or extant knowledge in this domain? Why focus on these 11 papers? Why focus on PLOS ONE? - It would be helpful to define acronyms e.g., what is a “6 ECTS course”? - This is unclear and perhaps needs rewording: “For problems with implementation specifics for methods described in the papers” - “RequirementR.3 is important to be able to contact the authors.” – this appears to just be a restatement of the requirement rather than a justification for including it. - To ensure the reproducibility of their own analyses, the authors may wish to consider saving and sharing the computational environment in which the analyses were performed. Various tools are available to achieve this e.g., Docker, Binder, Code Ocean. Reviewer #2: What did they do The authors tried to reproduce 11 statistical analyses published in PLOS ONE that used longitudinal data. This was done by cleverly making use of student labor in the light of a university course. For each paper, a detailed summary file on the OSF describes the study, the model, the analyses, and potential deviations in results. Those files further contain the used R code allowing the verification of this reproducibility study (Personally, I did not make use of that possibility!). General remarks I believe this work to be an important contribution to open science and a service to the scientific community in general. The manuscript is well-written and the authors delightfully refrained from being unnecessarily complicated. To put this work into perspective with similar empirical work on reproducibility in psychology, I suggest giving a more detailed description of methods, results, and, implications in the main manuscript. As of now, I am not sure which conclusions to reach about the state of reproducibility in PLOS one. A more detailed summary of the findings is particularly important in this case because each summary was written by a different teams of students making it very time-consuming to extract all the important information. Major remarks • I am confused as to the nature of this manuscript. Does it constitute a registered report? If so, the manuscript should clearly indicate what part of the work was done prior to the submission of the registered report and what was done afterward. • I am missing a (short) Method section where you describe the timeline of the conducted reproductions (when where authors contacted to provide code of analysis?, how did the students work on the assignment?, how (much) assistance did they receive from the teaching team?) In line 340 an internal peer review is mentioned – please provide more information on that. • I agree with what is being said in lines 208-212, however, I would like to have precise information about when the magnitude of the effect is the same. Further, the possibility of achieving the same numbers by deliberately deviating from the method description of the paper should be discussed as this has implications on the implied interpretations. • Roughly, we can group reproduction failures into 3 groups: Reporting errors in the paper, insufficient/incorrect description of methods or data that prevent reproduction, and software/algorithm differences. I would like to know for each reproduction failure the group to which it likely belongs. Since you exclusively used R in your reanalyses whereas only 1 of the 11 papers did so, I think it is important to discuss software differences (including differences in algorithms, default/starting values..) in detail. Whereas software differences are negligible for simple designs as ANOVA and t-tests, this cannot just be assumed for GEEs and GLMMs. A discussion of software differences is, for example, important to interpret the results for paper #1 (lines 225 to 234) and also line 257. Looking at your summary file of this paper “essay_01.pdf” it turns out that you have deviated in multiple ways for a multitude of reasons from what was described in the paper. As a result, it is hard to judge whether the original analysis contains reporting errors or not. your analysis of paper #1 A related issue is when you apply a different optimization algorithm. It might be of interest to try to reproduce those papers where the reproduction attempt was unsuccessful (and where the provided data does not seem to be the culprit) via the software package (and the functions therein) used by the respective authors. • 233 – If you believe that your R code does not converge properly, it should be changed until it does, no? If you are unable to fit the model in R, it cannot be judged whether the published results are approximately correct or not. Now, all we know is that the students assigned to this paper were unable to properly fit the statistical model to the data via R. • 316 – I would mention that in the abstract. The current abstract might give an incorrect impression as it is nowhere mentioned that the stated results involved author assistance (ideally, reproducibility in an open data journal should be possible without contacting the authors!) Minor remarks • 4 – use reproduction instead of replication. More generally, I suggest to use reproduce/reproducibility to describe computational reproducibility and replicated/replicability for new studies involving different data as this terminology is most commonly used in Psychology nowadays. • 12 - Longitudinal data includes variables that are measured repeatedly over time but those variables do not necessarily have to do with humans. • 111 – I would choose a more descriptive figure caption. • 112 – I would refer to R.1 in singular (i.e. requirement R1) • 130 – Is there the possibility of including additional papers? If so, I would like to see reproduction attempts of the 2 papers were the authors “prohibited” the use of their work to be reproduced. • 139-143 – Did you try to reproduce ALL figures and numbers reported in the paper that were related to the longitudinal study? If not, what was omitted and why? Please add 1 or 2 clarification sentences. • 144 – Reproducing someone else’s analysis typically involves many RDFs, yes. But, it does not make sense to say (line 145) that there were many decisions that would adhere to YOUR steps 1-4. Instead, you should write that there are multiple ways to read in/process/analyze the provided data that are not in disagreement with what is stated in the paper or the supplementary material. • 147 – Please be more specific. • 153 – The title is not self-explanatory, especially because it is written in the present tense. Maybe “Which statistical methods were used by the papers” instead. • 162 – “according to statisticians” I would refrain from using such a phrase. Instead, just cite relevant papers arguing for GLMMs over GEEs and, potentially, summarize some of their advantages. • 172 – See comments for 153 above • 184 – Did you always ask the authors for their source code. If not, when (before or after the 1st reproduction attempt?) did you ask for it. You provide some information in lines 201 to 207, but, I would like to know the specific time-line, and I want to know what was planned a-priori and what was ad hoc. • 247 – Where is the search procedure mentioned? • I would like to see the implications of the non-reproducible findings discussed. How many unreasonable original analyses (& conclusions drawn from it) could be identified? I know that this type of finger-pointing is uncomfortable, especially since only work from authors that provided both their data and responded to your e-mails were included in your sample, yet, it is important to estimate the rate of reporting errors and irreproducible findings in Psychology. Comments about Review and response to the registered report • The updated outline of the aim of this study is “Our aim with this study is to better understand the current practices in 11 PLOS ONE papers dealing with longitudinal data in terms of methodology applied but also in how results were computed and how it is made available for readers.” I find this unnecessary complicated and, more importantly, it does not reflect the content of your study well at all. Wasn’t the aim of this study to assess the extent to which papers analyzing longitudinal data in PLOS ONE could be reproduced by independent researchers. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: Yes: Richard Artner [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
Revision 1 |
PONE-D-20-25993R1 A computational reproducibility study of PLOS ONE articles featuring longitudinal data analyses PLOS ONE Dear Dr. Seibold, Thank you for revising your manuscript submitted for publication in PLOS ONE. Please accept my apologies for the relatively slow processing of your revision caused by several factors including me being on parental leave around the birth of our daughter and me having to homeschool her three proud brothers during the lockdown . The remaining reviewer (Reviewer 1 was unfortunately no longer available) and I agree that you responded very well to the issues raised in the earlier round and that your submission is very close to being publishable in PLOS ONE. The reviewer raises some minor issues that can be readily dealt with in the revision or responded to in your letter (in case you choose not to follow the suggestion). Also, please consider adding some references to some relevant recent studies on reproducibility and sharing of syntax and computer code and update the references referring to pre-prints that have appeared in the meantime. If you respond well to the remaining minor issues, I except to make a quick decision on your manuscript without needing to resend it out for review. I am looking forward to seeing your rigorous and interesting work appearing in print. Please submit your revised manuscript by Apr 23 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols We look forward to receiving your revised manuscript. Kind regards, Jelte M. Wicherts Academic Editor PLOS ONE Journal Requirements: Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #2: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #2: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #2: I Don't Know ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #2: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #2: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #2: I find that this revised paper well-structured and easy to read, and I can recommend its publication as new insights on reproducibility are much-needed. What I would like to see added is a more thorough discussion of the state of reproducibility in PLOS ONE (as well as more generally in psychology) together with its implications (both in the introduction and the discussion section). In particular, this paper should include references to all relevant literature on this topic (see minor remarks below). Otherwise, a reader of this paper will not be made aware of other important empirical findings on this topic. Below some minor remarks hopefully further improve the quality of this paper. • Abstract: The last sentence only states quite obvious things. I would prefer to read about non-obvious insights gained in light of this study. • Line 4: please include important recent studies on reproducibility such as o Artner, R., Verliefde, T., Steegen, S., Gomes, S., Traets, F., Tuerlinckx, F., & Vanpaemel, W. (2020). The reproducibility of statistical results in psychological research: An investigation using unpublished raw data. Psychological Methods. o Maassen, E., van Assen, M. A., Nuijten, M. B., Olsson-Collentine, A., & Wicherts, J. M. (2020). Reproducibility of individual effect sizes in meta-analyses in psychology. PloS one, 15(5), e0233107. o Obels, P., Lakens, D., Coles, N. A., Gottfried, J., & Green, S. A. (2020). Analysis of open data and computational reproducibility in registered reports in psychology. Advances in Methods and Practices in Psychological Science, 3(2), 229-237. • Line 33: What did other empirical investigations on reproducibility find? • Lines 160-162: I would not label a lack of knowledge of the exact calculations performed by the authors due to a lack of information provided as RDFs. RDFs are what the original authors had! Maybe you can write: „The description about all these steps was generelly vague (see classification of reported results in Artner et. al, 2020) meaning that there were multiple ways in line with the descriptions in the originial paper. This study, thus, exposed a large amount of Researcher degrees of freedom [23] coupled with a lack in transparancy about in the original studies.“ • I find the style in which the results section is written weird (until line 251). Why not just describe the results with a reference to the respective table. Now we have Tables first and it is not really clear if the text re-iterates the information in the tables or if additional information is provided. Also, why not merge tables 1, 2 and 3? Table 1 alone does not provide enough information to be included in the main text. Maybe the results section can be structured as follows: 1 • Line 239: When was the magnitude considered to be the same? It is important to exactly describe your criteria here to allow the reader to gauge the overall results of your study. Without knowing whether your criterion was lenient or rather strict, it cannot be done as each and every one of us uses individual metrics to gauge similarity. • Fig 2: General comment - Without knowing the range of parameter values it is hard to interpret differences in original and reproduced results. Why did you choose to report on the estimates of this article. Why not report, for example, on the results of article [1] instead? • Lines 261-264: Large differences in magnitude should result in different interpretation even in case of equal signs!!!!! • Line 299: substantial! • Line 351: Nevertheless we were only able …. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #2: Yes: Richard Artner [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
Revision 2 |
A computational reproducibility study of PLOS ONE articles featuring longitudinal data analyses PONE-D-20-25993R2 Dear Dr. Seibold, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Jelte M. Wicherts Academic Editor PLOS ONE |
Formally Accepted |
PONE-D-20-25993R2 A computational reproducibility study of PLOS ONE articles featuring longitudinal data analyses Dear Dr. Seibold: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Jelte M. Wicherts Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .