Peer Review History
| Original SubmissionSeptember 20, 2021 |
|---|
|
PONE-D-21-30265Forecasts of weekly incident and cumulative COVID-19 mortality in the United States: A comparison of combining methodsPLOS ONE Dear Dr. Taylor, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. ==============================All three reviewers have recognized the importance and timeliness of the topic. However, they have also highlighted several criticalities. Please refer to their detailed reviews for indications. Please be sure to answer all their comments in your revision.============================== Please submit your revised manuscript by Mar 03 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Maurizio Naldi Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match. When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes Reviewer #3: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: I Don't Know Reviewer #3: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: No Reviewer #2: No Reviewer #3: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The manuscript addresses an interesting and timely topic. The use of ensamble approaches is sound and comparisons across methods is very useful in practice. Some comments follow. 1. I would emphasize the importance of quantifying the uncertainty associated with forecasts. Therefore I kindly ask to report more predictive quantiles (1%, 2.5%, 5%, 10%, . . . , 90%, 95%, 97.5%, 99%) in addition to their point forecasts. This motivates considering both forecasts of cumulative and incident quantities, as predictive quantiles for these generally cannot be translated from one scale to the other. 2. The use of the interval score (Gneting and Raftery, 2007) is sound. The three summands can be interpreted as a measure of sharpness and penalties for under- and overprediction, respectively. I strongly believe that the weighted interval score (Bracher, J., Ray, E. L., Gneiting, T., and Reich, N. G. (2020a). Evaluating epidemic forecasts in an interval format. PLOS Computational Biology) shoul be considered. It combines the absolute error of the predictive median and the interval scores achieved for the nominal levels. It is a well-known quantile-based approximation of the continuous ranked probability score. 3. Maybe I miss something, or there is something swept under the carpet. In any ensamble-based approach, the choice of the weights is crucial. Please, provide more details on this point. 4. It would also be interesting to see a discussion of the performance of the models belonging to different families. Ioannidis et al (Ioannidis JPA, Cripps S, Tanner MA. Forecasting for COVID-19 has failed. Int J Forecast. 2020. http://www.sciencedirect.com/science/article/pii/S016920%7020301199) discussed that some approaches, mainly those referring to the SIR family, have failed in providing reasoable forecasts. On the other hand, data-driven approaches show much better performances in forecasting the evolution of the epidemic (see e.g. Mingione, M., Di Loro, P. A., Farcomeni, A., Divino, F., Lovison, G., Maruotti, A., & Lasinio, G. J. (2021). Spatio-temporal modelling of COVID-19 incident cases using Richards’ curve: An application to the Italian regions. Spatial Statistics, 100544; Girardi, P., Greco, L., & Ventura, L. (2021). Misspecified modeling of subsequent waves during COVID‐19 outbreak: a change‐point growth model. Biometrical Journal). Reviewer #2: Review of “Forecasts of weekly incident and cumulative COVID-19 mortality in the United States: A comparison of combining methods” Thank you for the opportunity to review this interesting paper. The topic is important. This analysis builds on a previously published study of combining methods for COVID-19 forecasting models in the United States and extends it using more data. The methods are interesting and seem mostly technically sound to me, although I think a comparison with percent errors is essential. I have a number of clarifying questions about the text. My main comment is about data and code availability. Pointing to the input data sources is not sufficient, in my opinion, as the authors must transform and manipulate the data to conduct the analysis. The devil is in the details with forecasting and predictive validity. The authors should publish their code and data in a public repository to facilitate a complete review of the research. That is the standard in the field for this kind of work (all other papers I’ve read on the topic have done so) and also required by the journal for publication: https://journals.plos.org/plosone/s/materials-software-and-code-sharing. Comments: Abstract 1. Methods section is vague. How specifically was the analysis conducted? “comparing accuracy” is not specific. Which metrics were used? How were they calculated? How did you design your holdouts? The basic details should be clear in the abstract. 2. The results section is also vague. Specific values should be cited. By what margin much did the best performing models overtake the others? 3. Conclusions: please indicate how this study did or did not concur with the previous study. What more can we learn from adding additional data? Methods 4. Why was only 1 year of data used when there are nearly 2 years of data available now since the start of the pandemic? 5. I am not sure that evaluating absolute error makes the most sense, as it biases the results towards higher mortality moments and locations. I would like to see a comparison to relative error metrics, such as the median absolute percent error. This has also been done in prior analyses so would be standard. 6. I would recommend placing the trimming and other details of how forecasts are combined in the main text, given journal length limits, and how central those details are to the work at hand. Results 7. How did the methods you describe here perform in comparison to the forecasts hub’s own internal ensemble model? Please highlight this important comparison. Reviewer #3: Thank you for the opportunity to review Taylor and Taylor’s manuscript that compares different methods of combining COVID-19 mortality forecasts. Their research is an important contribution to the forecasting field. Analysis of ensembling approaches is sorely needed; and it is also timely as the COVID-19 epidemic continues to change and forecasts continue to be used for operational decision making. Moreover, but as the field continues to move towards open and collaborative science, their work has direct application to future forecasting. Their clearly written manuscript has many strengths – it compares simple and more complex methods; it harnesses a year’s worth of data for evaluation; and focuses analyses on probabilistic distributions. In addition, there are also opportunities to improve to the text and analysis (see comments below) . MAJOR COMMENTS Abstract Line 30: I suggest cutting the text about “extended and new datasets”. It’s not apparently clear to readers what the authors are referring to if they haven’t read their previous manuscript. Line 35: The COVID-19 Forecast Hub collects probability distributions as well as point forecasts. The authors are referring here to the 50% quantile of probability distributions as ‘point forecasts’. Please refer to them throughout the manuscript as the 50% quantile, rather than a point forecast. [Also, see comments under Material and Methods, as I recommend excluding all 50% quantile analyses. These forecasts are 1) not useful for outbreak response, and 2) misleading in terms of communicating uncertainty.] Lines 35-37: Please list the evaluation metrics (mean interval score) and the combination methods here. Line 38 and 39: The first sentence of the Results is very confusing. Does the ‘average performance of these models’ refer to the Mean method, the Ensemble model, or something else? Line 46: How do you define “sufficient”? Length of historical data was beyond the scope of the analysis, but the text implies this was considered within the analysis. Introduction Line 89: Recommend not referring to reported data as ‘delayed’. ‘Reporting patterns’ is more accurate to describe the nature of the changes to the CSSE datasets and more accurately describes the descriptive nature of this part of the analysis. Materials and Methods Lines 109-110: The number of the week’s is very confusing here. What is the first week of the epidemic, and are you referring to the first week in the US? Because forecasts were not collected at ‘week 1’ of outbreak, I suggest starting the start date and the end date, with the start date referred to as ‘week 1 of the analysis’. Lines 113-116: I strongly recommend changing the inclusion criteria for this analysis. Models that were not included in the COVID-19 Forecast Hub should also be excluded here. This will help with comparability between the ensemble approaches. The COVID-19 Forecast Hub excludes forecasts that are improbable, such as if the number of cases or deaths exceeds the population size, not based on their predictions being ‘too large’. Details are provided here: https://covid19forecasthub.org/doc/ensemble/ Line 122: In the evaluation, did you use the CSSE reported counts at the end of the analysis period? OR was each date evaluated against the data available at the time. Please clarify in the methods. Line 122: What is the rationale for focusing only on the 95% PIs? Information can be gained be examining all intervals (7 available) and weighting them (a method applied by Cramer et al, 2021: https://www.medrxiv.org/content/10.1101/2021.02.03.21250974v3) Line 122: Because the interval score is a combination of calibration and sharpness, I wonder if the authors considered presenting all three metrics – calibration, sharpness, and IS? This might provide additional insights and if the authors are only presenting metrics at alpha of 0.05, they have the space to do so. Line 130: The field is moving away from communication of single numbers and towards ranges. Point predications are rarely used to communicate forecasts in the COVID-19 pandemic, and generally discouraged. Thus, I don’t find the point forecast analysis to be useful and suggest removing it. Line 142-146: Changes in the reported death counts were not always due to reporting delays. For example, the large spike observed in the winter 2020/2021 in Ohio reflects a change in how the state defined a death. Even the everchanging landscape of the pandemic, I’d recommend referring to these anomalies as “reporting patterns”, and perhaps defining examples of backlogged deaths, reporting dumps, or changes in definitions. Lines 157-159: Please define ‘overconfidence’ and ‘underconfidence’ and describe how they relate to the various trimming methods. Line 161: Can you say more about the ‘previous best’? It’s not clear to me why you added this model or if reference 42 is the correct reference for is. What does this add to the analysis? Line 171-172: Inclusion of individual models aids in the overall point that combining is better, however, the inclusion criteria is pretty strict here. Several models were consistent submitters to the COVID-19 Forecast Hub but missed a week here or there. Consider broadening this inclusion criteria. Line 171-172: I’d like to confirm that COVIDhub Baseline model was not included in this set? It’s not designed to be a true forecast but is rather a comparator point for the submitted models. Results Line 190: What were the thresholds used for the categories? How many states were in each category? Please also add to the discussion the limitation of not including a time component to the analysis, as the US experienced spatial heterogeneity in the outbreak and even lower incidence states had peaks, when model performance was subpar. Line 230: Please note which statistical test you are referring to here. Line 248: I think OK and WV are missing the dashed lines? If not, then there are no differences in reporting patterns overtime Line 255: Please present these data in either points or bard. Lines implies that the data are longitudinal. Line 266: Caution with describing Ohio and the individual models here. Because I don’t know which team model 33 is, I can’t speak to the accuracy of the text. It should be noted that several teams noted the spike in reported deaths in Ohio and assumed it to be an error, while other teams assumed it to be truth. Because this nuance is not available here, I recommend deleting mention of the individual models from the text. Lines 278-281: Can you share that sensitivity analysis as a supplement? Discussion Line 298: The main limitation of the analysis is the lack of temporal analysis. The epidemic varied over time and space in the US, and consequently, so did the forecast performance. While I do not think that the authors need to include temporal analysis, they should include this as a limitation in the Discussion. Line 302: As written, this implies that forecast type and timing was assessed in the analysis. Please provide a citation since this was beyond the scope of the manuscript. Line 337: Same comments as line 302. Please reference. MINOR COMMENTS Line 61: Source 15 has been published; please update the reference. Line 73: Center, in CDC, is spelled incorrectly. Line 93 and 98: Reference 33 is incorrect. It should be reference 26 here, as reference 33 refers to the reported data, not the forecast data or the collaborations surrounding the forecast data. Line 130 and 131: The sentence about point forecasts should be a new paragraph ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No Reviewer #3: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 1 |
|
Interval forecasts of weekly incident and cumulative COVID-19 mortality in the United States: A comparison of combining methods PONE-D-21-30265R1 Dear Dr. Taylor, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Maurizio Naldi Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #2: All comments have been addressed Reviewer #3: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #2: (No Response) Reviewer #3: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #2: (No Response) Reviewer #3: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #2: (No Response) Reviewer #3: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #2: (No Response) Reviewer #3: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #2: (No Response) Reviewer #3: Thank you to the authors for addressing all comments. In particular, the adjustments to include all quantiles, and thus better address uncertainty, have improved the manuscript. I have one very minor comment: The authors note that the COVID-19 Hub Ensemble method was switched to a median in July 2020. While this is correct, there was one additional change to the methods in November 20201. As of November 2021, the ensemble used a weighted approach based on WIS in 12 prior weeks (https://covid19forecasthub.org/doc/ensemble/). This is worth noting in the text. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #2: No Reviewer #3: No |
| Formally Accepted |
|
PONE-D-21-30265R1 Interval forecasts of weekly incident and cumulative COVID-19 mortality in the United States: A comparison of combining methods Dear Dr. Taylor: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Professor Maurizio Naldi Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .