Peer Review History
| Original SubmissionAugust 9, 2019 |
|---|
|
PONE-D-19-22557 Forecasting severe grape downy mildew attacks using machine learning PLOS ONE Dear Dr. Mathilde Chen, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Both Reviewers indicate the need to improve the paper significantly. We would appreciate receiving your revised manuscript by Jan 26 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols Please include the following items when submitting your revised manuscript:
Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. We look forward to receiving your revised manuscript. Kind regards, Andrea Luvisi Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf Additional Editor Comments (if provided): [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: No ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: No Reviewer #2: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: This manuscript explores a machine-learning model for predicting Grape Downy Mildew incidence and severity with the hope that growers can use this information to reduce the number of fungicide applications. After testing several models and algorithms, the authors find that date of disease onset is the best predictor of GDM incidence and severity, along with some climatic factors. The authors present an in-depth look at constructing a model, but the practicality of the model they have produced is lost. They acknowledge that growers will most likely not adopt practices involving increased scouting or delayed spraying. Spelling and grammar need to be corrected throughout. 113 and 114: Define site-year 115: 3 plants is very few. Is this data reliable? That is a very large range in plot size. 121: 1 to 57 is a very large range in number of site visits. Is this data reliable? 128: What do you mean by vine stocks? 147: Mislabeled as Fig. 1 226: Mislabeled as Fig. 2 272: Use of the term random forest twice 272: Mislabeled as Fig. 3 286: Mislabeled as Fig. 4 377: Use of attack is unusual – perhaps infection? 441: How this would reduce treatments by more than 50% is still unclear. This is a bold claim and I feel that it requires more detail so that readers can understand how the authors reached this number. It would be helpful to see an additional paragraph on how this number was derived. Fig 1. Caption does not mention incidence and does not offer enough detail on analysis. Medium severity looks very close to 0 – figures should be formatted to better show the severity instead of using the same scale as incidence. There is no mention of severity in the text, only incidence (Line 126) – describe the results in the text. Fig 2. Put labels in English. Lines are not distinguishable. What does this figure contribute to the manuscript? Fig. 6. Provide details of statistical analysis. More detail required. Fig. 7. This caption requires significantly more detail. Fig. 9. This caption and figure are unclear to me yet are the basis for the claim that fungicide applications can be reduced by 50%. More detail and explanation is required so that we understand this claim fully, because it has big implications. Reviewer #2: The authors study the ability of machine learning algorithms to predict the onset of GDM to reduce the number of fungicide treatments. The analysis is performed on an extensive data set spanning many years from the Bordeaux region. The problem is well motivated. The English is very good with only a few very minor mistakes. But, I have some open questions about the technical/ML approach. I believe the paper is at least a major revision. Major: Novelty is an important part of the publication process in general. Can the authors specifically comment on the novelty of this work? This is especially important because there are some cited works (36-39) that seem similar. It should not be up to the readers to infer the novelty of the work. How is the GDM onset date encoded as an input/feature? It would be a date, which is a character string, it’s not obvious how a study would numerical-ize this for a machine learning problem, and the authors should state this. I’m confused by the choice of ROC-AUC as a metric for measuring performance. The problem is a regression problem, and AUC is typically used for classification problems. Traditionally variations of squared error are used with regression problems. Is there a literature basis for using AUC for a regression problem? Currently, the set up for the experiment is this: 1) Regression analysis 2) Cast the problem as a classification problem by thresholding based on the median 3) Get AUC/ROC Why do regression in the first place if this was the end goal? I’m not sure I agree on thresholding the output based on greater/less than the median. This would only detect if a system is over/underpredicting GDM in a magnitude-less way. Isn’t the goal to get an accurate probability? You cannot measure error well if you do this. Can the authors elaborate one Line 186 what they mean by ‘median’ here? Is this the median value of the plot as a whole, median of the year, etc.? The outputs (incidence and severity) seem subjective because an expert would have to eyeball the spread of disease. I’m surprise the authors continued with some plots that had only 1 data inspection. The abstract states the authors use a year-by-year cross validation, this isn't a conventional cross-validation method and needs to be explained. Yet, it is not in the methods section. Very minor: Methods section: It’s conventional to state the number of input values for your data set, the readers should not infer this from a later figure. Line 53: I think this should be “Leaf” and not “Leaves” Line 83: Is mechanistic models the right term for this? Probabalistic models still need to determine some function of a model. Fig. 1: Box plots are helpful, but the authors should state the boundaries of the box. Is this the percentile? If so please state what they are. Figure 2: Axes are in French. Line 175: Technically, 100 trees is a parameter of the algorithm, this statement is not true in general. Line 232: Sever attack -> severe attack ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: Yes: Alberto C. Cruz, Ph.D. [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 1 |
|
Forecasting severe grape downy mildew attacks using machine learning PONE-D-19-22557R1 Dear Dr. Chen We are pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it complies with all outstanding technical requirements. Within one week, you will receive an e-mail containing information on the amendments required prior to publication. When all required modifications have been addressed, you will receive a formal acceptance letter and your manuscript will proceed to our production department and be scheduled for publication. Shortly after the formal acceptance letter is sent, an invoice for payment will follow. To ensure an efficient production and billing process, please log into Editorial Manager at https://www.editorialmanager.com/pone/, click the "Update My Information" link at the top of the page, and update your user information. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, you must inform our press team as soon as possible and no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. With kind regards, Andrea Luvisi Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed Reviewer #2: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Line 130-131 and in Figure 3: The revised manuscript still uses the term ‘grape stock’, which is not a commonly used term in viticulture. Are the authors = referring to trunks? Canes? Shoots? Reviewer #2: The manuscript now clearly states the novelty, and it is an impactful study on the ability to advise growers on fungicide treatments. The greatest strength of the manuscript is it’s discussion—rarely do statistical/machine learning works go into depth about “why” the model works. I have a few minor points, I believe the manuscript should be accepted (and no more than a minor revision). It’s clear now that the problem is a binary classification task. The AUC is indeed appropriate for this, but a statistical/machine learning problem *must* provide more than just the AUC. With R, it should be trivial to provide the additional metrics: true positive rate, false positive rate, positive predictive value, F1 score, confusion matrix, etc. (average over year-folds?). In particular, positive predictive value will indicate the promise of this work as a diagnostic tool, whereas AUC is more of a measurement of classification performance. Is it possible to also provide the ROC graphs for each method as well? Can the authors explicitly provide the a-priori rates for the classification task (preferably year-by-year)? Line 149: How did the authors chose a 1% infection rate? Is this based on some prior work, or is it a parameter? ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: Yes: Alberto C Cruz |
| Formally Accepted |
|
PONE-D-19-22557R1 Forecasting severe grape downy mildew attacks using machine learning Dear Dr. Chen: I am pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. For any other questions or concerns, please email plosone@plos.org. Thank you for submitting your work to PLOS ONE. With kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Andrea Luvisi Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .