Peer Review History
| Original SubmissionApril 5, 2021 |
|---|
|
Transfer Alert
This paper was transferred from another journal. As a result, its full editorial history (including decision letters, peer reviews and author responses) may not be present.
PONE-D-21-11098 Machine Learning Accurately Predicts the Multivariate Performance Phenotype from Morphology in Lizards PLOS ONE Dear Dr. Lailvaux, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Jul 23 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Christopher Nice, Ph.D. Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. Thank you for stating the following financial disclosure: The author(s) received no specific funding for this work. At this time, please address the following queries: a) Please clarify the sources of funding (financial or material support) for your study. List the grants or organizations that supported your study, including funding received from your institution. b) State what role the funders took in the study. If the funders had no role in your study, please state: “The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.” c) If any authors received a salary from any of your funders, please state which authors and which funders. d) If you did not receive any funding for this study, please state: “The authors received no specific funding for this work.” Please include your amended statements within your cover letter; we will change the online submission form on your behalf. Additional Editor Comments (if provided): Due to unavailability of reviewers for this manuscript, I am providing my own review to augment reviewer 1's comments. In any revision, please address both sets of comments. AE comments: This new approach for imputing missing data is well-presented and should be of high interest to readers involved with phenomics. My only real complaints are 1) that more details of the data used to train the model be presented and 2) that the authors are too succinct - some expansion of ideas and statistical issues or limitations would add value to this manuscript. Below are some elaborations on this theme and minor comments. 146: why "MVPpred"? The authors might justify the name of their approach. 153: change semicolons to commas 195-203: this needs some clarification, I think, with respect to the value of K. Was K=5 the initial approach, but then, given extensive missing data, the range of K was expanded (3-200) and the appropriate K chosen on the basis of MAE? Or is it RMSE? The supplementary figures (using MAE) do not comport with table 1. As an arbitrary example, for sprint speed, K=4 based on MAE from Fig. S6 but is reported in Table 1 as 84. Tables 3 and 4 should be switched in order - the stacking combinations (methods and currently table 4) should precede the results (table 3). "PCC", "MAE", "BAse Layer" and "Meta Layer" should be defined in the legends. Presentation of results: overall, this seems overly succinct and these three short paragraphs do not do justice for the work the authors have done here. For example, in the regression results, jump power is reported because of highest predictive power, and jump acceleration is mentioned at the end (250) but what about overall conclusions? Range of results. In other words, it is difficult to comprehend what the authors' main point is here. In the next section on stacking results (257) do the authors mean that SM2 outperformed the others in ALL cases as indicated by Table 5? No, according to the supplementary tables, the text is correct and the table is mistaken. I think readers would benefit from a fuller exploration of the results. Discussion: as above with the results, the discussion could be broadened to help readers appreciate the full scope. Yes, MVPpred should not be a replacement for collecting data, but, as the authors state in the introduction and the first sentence of the Discussion, such data is logistically difficult to come by, or impossible. This model seems to be a valuable tool as a consequence. But readers might appreciate a discussion of the details. How and why are stacking methods improving results? Why is the SM1 configuration superior in all (or most) features? Are there limitations (beyond inferring causality), or inappropriate uses of this approach? [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: No ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: No ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: No ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: This manuscript describes a model for predicting performance of lizards from morphological measurements. It makes two claims: one is that the model is extremely accurate at predicting these performances, and the second is that this is related to phenomics: high-throughput, multivariate phenotyping. The first claim is not supported by the kind of evidence one needs to evaluate it. The second is a non-sequitur. If you don’t measure the phenotype, you are not phenotyping. A method which successfully predicts performance would be an extremely useful tool, and this software pipeline may be such a tool. The problem here is that the two measures reported here do not adequately summarize accuracy. High R^2 is certainly a good thing, but it is not enough without seeing the distribution of the data and the predictions used to validate the model. To take a really simple example, if we have two clusters of jump performances, with one cluster being can’t jump, and the other similar in jumping ability, you can get a very high R^2 without the model being accurate: all it has to do is say one is low, and the other is high. The MAE values are essentially uninterpretable without knowing the mean performance and the units in which they are measured. The proudly reported value of MAE of 1.48 has NO units attached!!! Proportional error of the predictions would be a very meaningful statistic. There are excellent biological reasons to doubt that the repeatabilities of these performance characteristics could be high enough to produce the R^2 values reported. If you run or jump the same lizard and different equations, I would be astonished if repeatabilities could ever approach 98%. The assay for sprint speed is descrbied as doen at an angle of 45 or less. How can it not matter whether you sprint uphill on on the flat? Since this is the case how could an individuals morphology ever predict performance with such high accuracy. In short, the manuscript makes an extraordinary claim, but fails to back it up with meaningful statistical evidence. What is absolutely essential to evaluate the accuracy is that we see the number of species actually measured in each case, as well as the error of each prediction when used in the test data set. Similarly, we really need to know whether the within-species variation is also explained by morphology, or whether this is just the mean that is explained. Since there is no enumeration of studies, I don’t know whether they are predicting the performance of 10 species from that of 58, or predicting 58 from 10 for each performance trait. What are the within-species sample sizes? How are those 2000 individual lizards divided among each trait, and species? Why is there no figure showing measurements and predictions with meaningfull errors for each? If this model is in fact highly accurate by some meaningful measurement, this would be a very important result. It has important implications for phenomics, suggesting the dimensionality of the phenotype as whole is not so very high. Rather than BEING phenomics, such predictive ability would suggest that we do not really need phenomics. Trying to present the results AS phenomics is misguided, but the implications FOR phenomics are very interesting. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: Yes: David Houle [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 1 |
|
PONE-D-21-11098R1Machine Learning Accurately Predicts the Multivariate Performance Phenotype from Morphology in LizardsPLOS ONE Dear Dr. Lailvaux, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. The reviewers found this revision to be improved and more clear. However, Reviewer 1 in particular is interested in more details regarding model validation / performance evaluation. Given that clarity is at a premium with the introduction of new methods, I ask that you consider these comments carefully. Please submit your revised manuscript by Nov 14 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Christopher Nice, Ph.D. Academic Editor PLOS ONE Journal Requirements: Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice. Additional Editor Comments (if provided): [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: (No Response) Reviewer #2: All comments have been addressed Reviewer #3: (No Response) ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes Reviewer #3: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: No Reviewer #2: I Don't Know Reviewer #3: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: No Reviewer #2: Yes Reviewer #3: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The alterations to the manuscript having improved its clarity on many basic points. However, now that I understand what the authors are doing better, it is clear that some of my original questions remain. Tthe authors appear to have misread my comments on measures of model accuracy. Because the evaluation of the accuracy of the model predictions is dubious in many respects, I am still worried that the authors give quite a misleading picture of their model’s performance. First, I want to be clear on a critical point, and that is whether the model is being asked to predict imputed data in the 10-fold cross-validation step. The authors describe imputation, and then fitting of the overall model to the imputed + observed data in some detail. However, the performance evaluation is simply described as splitting the data into subsets, followed by standard cross-validation. There are two potential issues here. I would like to assume that the authors removed the imputed performance values from the 1/10 of the data that is used as the test set. Please make explicit that this is not what you are doing. Only actually observed maximum performance values are legitimate to use in the test data set evaluate the accuracy of the model predictions. While it is fine to use imputed values if they improve model predictions, it is NOT a test of accuracy to use a test data set with imputations to see whether the imputations are predicted by the model. Statements like this one at line 297-298 make me afraid that the authors HAVE made this fundamental error “treadmill endurance was measured for only ~7.8% of individuals but inferred with an accuracy of 0.95 throughout the dataset.” If they have, then all measures of accuracy in this manuscript are completely bogus. The more general issue is that an overall measure of correlation or error is not very informative, particularly in a data set with massively unbalanced sample sizes. As the authors note the data set is heavily weighted to Anolis, and indeed 45% of the specimens are in just three species. This means that whatever individual-wise measure of accuracy is computed is mostly reflecting the model’s performance in those species with the largest sample sizes. For example authors say (lines 323-325) “making it all the more remarkable that our model was able to make accurate predictions even for sparsely sampled taxa.” The problem is that the authors have not calculated the accuracy of predictions on sparsely sampled taxa, just on the overall data set. To make such statements the authors need to report accuracy for each species. The authors single PCC or MAE value is effectively weighted by within species sample size. A more representative overall measure would be an unweighted mean (or median) PCC or MAE value over species. This would help get at another unaddressed issue from my previous review and that is the issue of proportional errors. While it is a great improvement that we now have the overall means of each performance measure, an MAE value might be very small for an organism with a high predicted value, and very large for an organism with a small predicted value. There is still no summary of the actual performance values to enable a reader to evaluate this issue, as the authors make no attempt to do so. The statement on line 295 about MAE is still made without units, and is completely meaningless without thinking in proportional terms. For example the extreme values cited 0.003 meters jump and 1.75 seconds are each close to a 1% error, and not really very different. But is the error low throughout the range of performance values? Interpret all MAE values proportionally. Another cross-validation that would also be informative is to do it species-wise – that is leave out each species from the training data set and ask how well its performance is predicted by the remaining species performance data. This would be an interesting test of the author’s contention that phylogeny does not matter much. If that is true then there should be little cost to cross-species predictions. If cross-species predictions do well within the data set, then this would suggest that the model may indeed be useful when applied to species whose performance has not been measured. I still argue that prediction is not really that relevant to phenomics, except in the way outlined in my previous review. As the lead author on the article cited, perhaps you should give my point of view some credence. Minor comments. Line 105 Cormorants dive from the water surface, not while in flight. Line 262-271. I believe the authors mean to refer readers to Tables S4 and S5 in this paragraph. Reviewer #2: Dear Editor, I have considered the revised manuscript entitled 'Machine Learning Accurately Predicts the Multivariate Performance Phenotype from morphology in Lizards’ by S. Lailvaux, although I did not review the original version. I also read carefully the authors’ extensive responses to the referees' comments and questions and paid special attention to how these comments were addressed (where necessary) in the manuscript (the added track-changes version was very helpful for this). In my opinion, the authors did this revision very thoroughly and respectfully. With regard to the content, I must admit that I am not at all familiar with the statistical (and computer) techniques applied to this massive lizard-dataset. On the other hand, I am sufficiently familiar with ecological/morphological research (in the evolutionary context) to see the enormous potential of this methodology. This is especially the merit, also for the 'mathematical layman', of the very comprehensible introduction and discussion. The only thing I cannot quite assess is what the direct applicability (and thus in a sense the valorisation value) of this method might be to other, new cases (e.g. a study of the link between morphology and performance traits in arthropods). Morphometric and performance data will always be needed to train the routine to make predictions. But how extensive should this training dataset be? How 'lizard' specific is the current procedure (in other words: is the protocol directly applicable to other systems)? Can the classical 'ecologist' apply this method autonomously ... or will the participation of colleagues from computer sciences be necessary? Etc. The authors may wish to provide a perspective in this respect in a short paragraph in the discussion. Reviewer #3: A concise report on the application of a statistical method to morphology>performance data as a way to deal with incomplete datasets. The paper addresses a real problem and provides a robust statistical solution. I've got no major criticisms of the methods or results. I have a few comments that might be considered to help improve, especially, the discussion: line 120: the model only predicts performance for incomplete performance datasets, correct? Or can the model work with incomplete morphological datasets too? Please check the wording here. line 144: ML. Every time I read ML I think maximum likelihood but it is machine learning. I might suggest avoiding this abbreviation altogether the paper already has a lot of abbreviations, so one fewer would make for a bit less mental work for the reader. line 297: I am having a hard time understanding how this is even possible. So if I measure endurance on 7.8% of my samples and then use the model to predict the other 92.2% of samples...how do I 'know' if that prediction is at all accurate? You don't really 'know' what those endurance values are, you only have a prediction based on other traits that is dependent upon very poor samples of 'known' values. Do the authors really believe that if I measured the other 92.2% of species that my measurements would fall within the prediction 95% of the time (or rather does the model tell us that)??? It seems like a huge leap of faith based on very weak underlying sampling. Line 341: I think you've missed a key idea. While endurance (i.e. the ability of muscles to sustain contraction to propel the animal forward at a given speed) is no doubt most closely linked to the cardiovascular and pulmonary systems....it is also TOTALLY dependent upon the limb morphology and body dimension of a given species. Shorter limbed species must cycle their limbs more often to maintain speed, thus taxing their muscles more than a longer limbed species that ran the same distance. Dachsunds will always tire before greyhounds and some (or even a lot) of that is related to their limb shape! Line 366: Ughh...you undercut one of the main benefits of your model...prediction. But it is true that applying this model to other species is iffy. You might consider some text here to explain what type of dataset might be needed to build a model that ultimately COULD be used across other species. Line 334: Feels like a bit of a cop out. I think there is more causality that you can infer here than you give yourself credit for. Or rather, there is more biology here than the paper currently digs into. I realize the point of the paper is to demonstrate and validate the statistical model....but it sure would have been nice to see a bit deeper dive into the biology of how these performance traits trade-off or facilitate, etc. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: Yes: David Houle Reviewer #2: No Reviewer #3: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 2 |
|
Machine Learning Accurately Predicts the Multivariate Performance Phenotype from Morphology in Lizards PONE-D-21-11098R2 Dear Dr. Lailvaux, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Christopher Nice, Ph.D. Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: |
| Formally Accepted |
|
PONE-D-21-11098R2 Machine Learning Accurately Predicts the Multivariate Performance Phenotype from Morphology in Lizards Dear Dr. Lailvaux: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Christopher Nice Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .