Peer Review History
| Original SubmissionFebruary 7, 2022 |
|---|
|
PONE-D-22-03759BlotIt - Optimal alignment of western blot and qPCR experimentsPLOS ONE Dear Dr. Kemmer, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. The three reviewers agree that the manuscript would be a valuable addition to the literature, provided that a number of aspects are clarified, and I concur with their assessment. In the review reports you will find a list of suggestions to clarify a number of technical aspects of the proposed method. Notably, two of them agree that it would be interesting to compare BlotIt with alternative approaches. As pointed out by Reviewer 2, one possibility would be to perform a comparison using simulated datasets. If possible, such a study would greatly enhance the contributions of the paper. Please submit your revised manuscript by Apr 16 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Alejandro Fernández Villaverde, Ph.D. Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match. When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section. 3. PLOS ONE now requires that authors provide the original uncropped and unadjusted images underlying all blot or gel results reported in a submission’s figures or Supporting Information files. This policy and the journal’s other requirements for blot/gel reporting and figure preparation are described in detail at https://journals.plos.org/plosone/s/figures#loc-blot-and-gel-reporting-requirements and https://journals.plos.org/plosone/s/figures#loc-preparing-figures-from-image-files. When you submit your revised manuscript, please ensure that your figures adhere fully to these guidelines and provide the original underlying images for all blot or gel data reported in your submission. See the following link for instructions on providing the original image data: https://journals.plos.org/plosone/s/figures#loc-original-images-for-blots-and-gels.
In your cover letter, please note whether your blot/gel image data are in Supporting Information or posted at a public data repository, provide the repository URL if relevant, and provide specific details as to which raw blot/gel images, if any, are not available. Email us at plosone@plos.org if you have any questions. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: BlotIt - Optimal alignment of western blot and qPCR experiments Summary The authors present an automated procedure for the normalisation of relative data, such as data obtained from Western blot or rt-qPCR. Relative data are usually not directly comparable across replicates, because of different arbitrary units obtained, and so require normalisation. Using likelihood-based optimisation, the proposed method is able to estimate how to scale each replicate, and at the same time provides an estimate of mean and variance for the underlying data. The approach is flexible in various way, as it allows different error models to be used, an also it is able to combine data from multiple blots as long as there is one experiment shared across pairs of blots. The work presented is certainly worthy of publication, although some clarifications and minor corrections are necessary before I can give my final approval. Major 1. Missing discussion of the fact that Yij at extremes of the dynamic range of detection can have relatively low signal to noise ratio or high residual error epsilon. While the error epsilon is initially introduced as potentially different for each Yij, it is then simplified to have a variance proportional to the value of the measurement (sigma_ij=e_rel*yi/sj). While this can be considered an appropriate generic error model, its drawbacks should also be highlighted, such as its underestimation of variance for relatively low intensity blot values (which notoriously have a low signal to noise ratio). It would also be more appropriate to adjust lines 80-83 page 3, to reflect the fact that this error model (Eq 2) efficacy depends on the assumptions chosen to implement h and whether the data follows these assumptions, rather than just describing it as a superior approach. The advantages and disadvantages of the chosen simplified model should be highlighted. I guess the advantage here is the reduction of parameters to estimate (just e_rel), while the disadvantage is the loss of model flexibility. For example, it will ignore that low intensity values might have a much higher relative variability (because of the low signal to noise ratio). 2. Please define a set or range of values for the indexes i and j. For example there could be I conditions and J blots with i in the range 1,…,I and j in the range 1,…,J. In principle, each equation should have a definition for i and j range, like ‘for all i in (…), j in (…)’. The simplest case would be when the set of experiments i is the same for all J. However, the authors imply that different j can have different sets of experiments, so different i? In this case perhaps it makes more sense to talk about indexes i that belong to sets Ij (that is indexed by j) and that the intersections of sets Ij need to be at least pairwise not empty (i.e. share at least one condition i). This might also affect the notation in equation 10. 3. How is the mean of the true values y dash (equation 10c) defined? This is particularly important if there are different conditions i on different blots j (like in Figure 2), I guess this will be a mean across all i regardless of what blot j they belong to? 4. In equation 7, page 4, the measurement error term is not shown. How do you expect the value of Ysij be determined if the error is unknown? For example, arranging Eq 1, Ysij = (Yij+eij)*(s hat j). Because the error is potentially different for each data point, how will the mean and sd of the aligned data be affected? More explanation is given later, but here a clarification and showing the particular example (f=Y/s) next the generic equation would help the flow. Perhaps, it would help to mention the ideas of error propagation here, with examples and simplifying assumptions. 5. At lines 129 to 131 page 4, the authors assert that for dynamic models parameter estimation mean and standard deviation is sometimes preferred, citing the case of low number of replicates. One could argue the opposite, that if there are few replicates, then all data should be used for fitting, because the average and sd alone may not properly represent a given datapoint. I suggest to remove or add more literature or explanation to support the authors claim. 6. The claim at lines 133 and 134 that the estimated errors are more reliable than the data spread obtained from replicates, should be further explained. For example, this could be true only if the data agree with the error models and there could be exceptions (see above for low signal to noise datapoints), and also it depends on how the data spread of replicates is calculated, for example if the replicates are done all on the same blot, then it might be more accurate, but if they are done on different blots, then the spread is completely dependent on the normalisation applied and possibly the value of other datapoints. 7. Why there is no error term in equation 8? Also, in this case it would be useful to give a generic idea of how the error can be inferred and follow Eq 8 with the concrete example f=Y/s, and simplifying assumptions. 8. On page 5 line 146, please reconsider and rephrase the claim that the optimal theta is obtained by minimising the spread of the replicates, because the likelihood model is not just about trying to reduce the spread (low sigma). 9. In the error determination section (page 6), equations 13 and 14 require a reference, please add. Also, it would be nice to accompany these equations with concrete examples for the specific cases and simplifying assumptions described (like f=Y/s). For example, sigma_s, if I am not mistaken, simplifies to sigma_s_ij=s_j*sigma_ij. 10. Perhaps it would be of interest adding to the discussion some speculative yet useful cases. For example, what happens when two blots share only one experiment, but in one blot j the measurement is likely to have a high signal to noise ratio, while in the other blot j’ the same experiment has a very poor signal to noise, perhaps because it is a low intensity value. Would the proposed model be able to propagate the variance of the normalised data? Would there be enough data to constrain the model optimisation? 11. What other error models are available in blotIt besides e_rel*value and e_abs? Minor: 12. Shouldn’t Equations 5 and 6 mirror the definitions in Equations 3 and 4? Equations 5 and 6 seem to be a mix of the generic vector format Equations 1 and 2 and equations with specific indexes i and j such as Eq 3 and 4. Please, choose one format for clarity, probably the format with the indexes i and j would be more suitable for Eq 5 and 6 following the flow of the paper. 13. Line 114, page 4, what is the dimensionality of y hat, s hat and e hat? Probably this will be clearer once the indexes are clarified (see above) 14. ‘Therefore’ might be more appropriate than ‘therefor’ at line 8 page 1, and also line 50 page 2, and also in other places across the manuscript. 15. Line 118 page 4, form -> from 16. In equation 7 page 4, the variable Y_s is undefined. I was initially confused by it because the text that precedes it talks about the true values y. It might help to write a sentence giving a proper definition for Y_s, and accompany this equation with another equation exemplifying what Y_s_ij look like for the simplified model f=Y/s and error e_rel*Yij/sj. Reviewer #2: This manuscript describes a normalization strategy for western blotting as well as many other assay types that can put relative data from different experiments or replicates on the same quantitative scale. This is needed often because experiment specific factors cause the scaling to not be comparable between replicates or different experiments. A main claimed novelty is that the same condition need not be contained in every experiment for normalization. Rather, each experiment needs to share at least one point with one other experiment. I would have liked to have seen more application and discussion to data sets that have such a feature, and showing where current methods fail. And perhaps some discussion of how often this scenario is found in the literature and would be needed. There are other points listed below that may be important to address: 1. What justification do the authors have for assuming normality in errors for western blots? How much does that affect the conclusions of the paper? Could alternative error models be used and BlotIt still functions well? How does that impact application to other data types? 2. It is appreciated that the authors technique is claimed to work when data points are not shared between all replicates and/or conditions. How common or rare this is for bench scientists performing replicates or experiments was not discussed. What novelties or advantages does blotit have when data points are shared among all replicates or experiments? Getting more clarity on that would help make the impact and uptake of the paper clearer. 3. How does the proposed approach differ from the scaling factor approach described here: 10.1038/msb.2009.4 ? There is a general lack of comparison to other analysis methods which have been established and used for quite a long time, as cited by the authors. 4. The discussion of how to use the R code and format it seems more appropriate for detailed methods section, not the results section. 5. In Figure 2, how do alternative methods for normalizing data compare to BlotIt? 6. Perhaps a simulated data study where data points are actually shared between all experiments, but are hidden to see how BlotIt does, could be an effective analysis to demonstrate usefulness and also compare to other normalization methods. 7. How would one compare the common scale data as the output of BlotIt to model simulations that would have a different scale (e.g. absolute concentrations)? Often comparison to and use with dynamical models is cited in the paper as a main motivator, but discussion with respect to this is lacking. 8. Therefor--> therefore Reviewer #3: Review for the manuscript "BlotIt - Optimal alignment of western blot and qPCR experiments": The manuscript proposes a novel alignment method for relative data. Via optimization, it finds a version of the data on a common scale. An implementation in R is provided. The manuscript is overall well written and easy to read, while in some places it could be more specific. In my opinion, the new method is interesting and will find usage, while it is maybe not a mayor conceptual breakthrough. I have a few comments, which I think should be addressed in a revision: Content ------- - it could be made clearer that, unlike e.g. the approach by Weber et al., the approach is essentially model-free, i.e. only dependent on the data and a noise model, but not e.g. a post-hoc employed ODE model. - A comparison to e.g. the method by Weber, as well as the method by Degasperi, in a situation where it is applicable, would be of interest, e.g. regarding predictions, efficiency and uncertainties. It is however understandable if this is beyond the scope of this work. - l. 35f: "disadvantage of enlarging the parameter space drastically": The approach by Weber (or also later papers by Loos et al. "Hierarchical optimization for the efficient parametrization of ODE models" and Schmiester et al. "Efficient parameterization of large-scale dynamic models based on relative measurements") appear to argue explicitly that the parameter space is effectively not enlarged by a hierarchical formulation. - l. 36f: I did not understand "estimates of the scaling parameters might be biased by the model equations[,] hampering hypothesis testing and therefor[e] interpretation". - l. 82: "the error model considers the variance information from all experimental data" as opposed to "calculation of errors based on the spread of measurement values" does not get clear to me. - l. 146: There appears to be a constant $\\pi$ missing in (10b), when deriving (10a) and (10b) from a normal density, which affects the relative impact of both terms. - l. 146: I think it could be clarified that $\\bar y$ in (10c) denotes the mean (over all data points?)? - l. 146: Where does the $10^{-3}$ come from? This appears to be rather arbitrary and may affect how much emphasis the method puts on normalization. How sensitive is the method with regard to it? Or can it be chosen in a problem-specific manner? Or would there be an alternative formulation as an optimization problem with explicit constraint $\\bar y = 1$? - l. 146: Can (10c) be interpreted stochastically (as (9) claims to describe a probability density)? - l. 158: "This drastically improves numerical stability": This is surely an accurate fact, yet a reference may be good. Is the method applicable to negative data? - l. 167ff: References on the parameter/data ratio, the variance underestimation, and the Bessel correction would be good. - l. 180: The confidence interval appears based on a local Taylor approximation given asymptotic normality of the maximum likelihood estimate (with covariance matrix given by the inverse Fisher information matrix). Conceptually, there should be alternative methods, e.g. based on Wilk's theorem or sampling. Maybe a contextualization would be good? - l. 184: What do the authors mean by the FIM is "represented by the Hessian"? - implementation of the method: How is the optimization problem solved? Are gradients available? Does the problem have multiple local optima? - implementation of the method: How computationally expensive is the method? Does it scale to e.g. aligning single-cell data, where normalization is often done simply be cell size? - As mentioned before, a comparison with alternative methods, and a discussion on how to use the scaled data in downstream analysis would be of interest, but it is understandable if this is beyond the scope of this work. A particular question that may come up is: E.g. an ODE model will output values on a certain scale, which may be different from the normalized scale by the presented method. Would this necessitate the use of scaling factors when fitting the ODE model still? Grammar ------- - e.g. l. 8, 37: While this word also exists, you probably mean "Therefore" in multiple places. - l. 37: "[,] hampering" - l. 195: "concentrations[,] meaning" - Table 1: comma in $Y_s = f^{-1}(Y, \\hat s)$ ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No Reviewer #3: Yes: Yannik Schälte [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 1 |
|
BlotIt - Optimal alignment of Western blot and qPCR experiments PONE-D-22-03759R1 Dear Dr. Kemmer, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Alejandro Fernández Villaverde, Ph.D. Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: (No Response) Reviewer #2: All comments have been addressed Reviewer #3: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: I would like to thank the authors for the additional work and for replying to my previous questions, which I consider answered. I just have a few additional point, just the first point is of major concern, and hopefully can be addressed easily. 1. I appreciate the additional simulation study, however I am concerned about the criteria chosen to evaluate and compare the methods. The authors write: “The performance of the individual methods was evaluated for each of the data realizations based on the spread of the scaled data”. Do the authors mean that the methods producing the normalised data with the narrowest standard deviation are preferrable? I think that one of the points of the mentioned Degasperi et al, was that data themselves have a spread, and that if we underestimate such spread this could also be problematic, for example making us believe that there is a difference between two conditions just because our assumptions have reduced the uncertainty of their mean value. So, I wonder whether the goal should be to prefer a method that produces a spread of the scaled data that is as close as possible to that of the simulated data. 2. The Methods section begins with the definition of the sets I and J, as well as measurements Yij. If I understand correctly, the key message here is that measurements Yij are comparable across the index i but not j. If so, this should be stated clearly, and some examples perhaps modified to avoid confusion. For example, the examples of biological effects include things that are comparable like different conditions and time points, but also things that are not comparable like different protein targets in a Western blot. 3. Line 99 of update text: “all experimental data, what allows for a reliable error”, change ‘what’ to ‘which’? Reviewer #2: The authors have done a reasonable job of addressing the concerns raised in the review. I think the paper should be published and will be of use to biological data analysis ideas. Reviewer #3: My comments to the first version have all been sufficeintly addressed, with some minor issues: - l. 425 and l. 446: "bloIt", and in a few other places. I guess the method goes only by "blotIt" - e.g. missing spaces and commas and shortforms like "didn't" in a few places in the newly added text - 10^{-3} in (10c): I would recommend to include the answer to the question as part of the manuscript or supplement. - same for the answer to how the optimization problem was solved ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No Reviewer #3: Yes: Yannik Schälte ********** |
| Formally Accepted |
|
PONE-D-22-03759R1 BlotIt - Optimal alignment of Western blot and qPCR experiments Dear Dr. Kemmer: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Alejandro Fernández Villaverde Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .