Peer Review History
| Original SubmissionJuly 6, 2023 |
|---|
|
Dear Dr Zhou, Thank you very much for submitting your manuscript "SPIN-CGNN: Improved fixed backbone protein design with contact map-based graph construction and contact graph neural network" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments. We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript will be sent to reviewers for further evaluation. When you are ready to resubmit, please upload the following: [1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. [2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file). Important additional instructions are given below your reviewer comments. Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts. Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments. Sincerely, Yang Zhang Guest Editor PLOS Computational Biology Nir Ben-Tal Section Editor PLOS Computational Biology *********************** Reviewer's Responses to Questions Comments to the Authors: Please note here if the review is uploaded as an attachment. Reviewer #1: The study by Zhang et al. presents a new deep learning method for fixed backbone protein sequence design, SPIN-CGNN. The overall architecture is similar to the latest graph-based neural network methods such as ProteinMPNN and PiFold. The major new advancements in the paper are adaptive construction of graph edges through specific distance thresholds, the use of selective kernel networks for node and edge updates, as well as the introduction of symmetric and second-order edge information. The work is nicely presented and should be suitable for publication after revisions are made. My specific comments may be found below. 1. I appreciate the effort the authors put into retraining ProteinMPNN and PiFold; however, the authors should compare the performance of these programs based on the publicly released, pretrained models by each respective lab. It is not the author’s responsibility to improve the pipeline or training procedure for the control methods. Moreover, others could argue that the results are worse due to the retraining and not because of methodological improvements by SPIN-CGNN. Thus, the results of the default models should be provided whenever possible. If redundancy between the training and test sets is an issue, the authors may examine the performance on a set of non-redundant targets. 2. The authors state on lines 506-507 in reference to Table 4, “we do not have the performance for all individual proteins for most methods.” Did the authors take the results on the test set from the papers directly? How do we know that the reported results are accurate? The authors should download each program and run them locally on their test set if possible to verify the performance of each method. This should be done for all results presented in the manuscript. 3. I would suggest that the authors add in a comparison with RosettaFixBB and/or RosettaFlexibleBB, which are probably two of the most widely-used energy function-based methods for protein sequence design. 4. What is the formula for calculating the LCR in Tables 5 and 6? 5. In Table 5, what is the sequence recovery rate in buried vs solvent exposed regions of the protein. Are the deep learning methods very good at recovering buried regions but not at recovering exposed regions? Do the physical energy-based methods outperform the deep learning methods on the exposed regions? 6. In Table 5, what is the BLOSUM score for OSCAR-design. Currently, the BLOSUM score is calculated based on the probability distribution similar to the perplexity, making it impossible to calculate this value for OSCAR-design. However, the authors may additionally calculate the BLOSUM score based on the actual designed sequences similar to the sequence recovery rate. Do the physics-based methods have more favorable BLOSUM scores than the deep learning methods? And if so, what does this mean? 7. In Table 6, what is the sequence recovery rate and AA composition deviation for each of the methods on the Hallucination129 test set? 8. In Tables 5 and 6, how was AlphaFold2 run? Were just the single designed sequences used or did the authors use MSAs? Also, were templates with >30% sequence identity to the queries excluded. If not, the TM-scores may be better for the deep learning methods as the native structures themselves could be identified as templates given the high sequence identity between the native and designed sequences. 9. The results in Tables 5 and 6 are very interesting. For example, the median sequence recovery rate for SPIN-CGNN is nearly 60% higher than the physical energy function-based method, OSCAR-design; however, OSCAR-design produces sequences with AA compositions closer to the native distributions. Does this mean that the OSCAR-design sequences are more physically realistic? Additionally, why are the results for the deep learning methods significantly better than the physical energy function-based methods on the native structures but worse on the designed structures? Are SPIN-CGNN and the other deep learning-based methods very good at recapitulating high-confidence, conserved regions in the protein structures but fail to recapitulate less conserved regions? What are the physical characteristics of the designs produced by the deep learning methods? Do they introduce steric clashes or physically unrealistic residues? I believe a closer examination of the physical quality and stability of the designs is warranted. The authors could run MD simulations to verify the stability of the designed sequences and provide a closer examination of the physical quality of the designs by examining such things as steric clashes and MolProbity scores, among others. Reviewer #2: Zhang et al. presented SPIN-CGNN as a graph neural network-based deep learning approach for de novo protein sequence on fixed protein backbones. Compared to the conventional GNN methods based on kNN graphs (where k is typically a value fixed to 30), SPIN-CGNN constructs graphs based on the distance-based protein contact maps, thus allowing an adaptive “k” for different local structures of the backbone. They also introduced symmetric and second order edge updates in their GNN. The approach, though lacking a wet-lab validation, is well benchmarked in silico with a variety of metrics, including native sequence recovery, perplexity, amino acid composition, and amino acid substitution, to just name a few. SPIN-CGNN achieved SOTA performance on almost all metrics, with the exception that the relatively high ratio of low-complexity regions in designer sequences compared to native sequences. I recommend publishing this method paper after a few minor comments are addressed. 1. In lines 200-202, the u_i and v_i should be defined in line 192. 2. Lines 423-424, “ProteinMPNN reimplemented model” is somewhat confusing. I found that ProteinMPP was also reimplemented in ref. 25 and the rate of 45.96% was from the reproduced version therein. Thus, I suggest the authors change “ProteinMPNN reimplemented model” to “ProteinMPNN implemented by Gao et al.” 3. In Table 2, “4.36±0.01” in the “Model 3” row should also be highlighed. 4. In Table 4, SPIN-CGNN outperformed “LM-DESIGN + PiFold” all categories of perplexity test and the “Short” and “Single-chain” categories of Median recovery (~7% and ~1%, respectively) but was ~1% worse in “All” of Median recovery. Could the authors compare the performance of SPIN-CGNN and “LM-DESIGN + PiFold” on the targets that are not categorized as “Short” or “Single-chain”, so that readers can understand how much worse SPIN-CGNN on these targets? 5. The concept of “low complexity region” should be clearly defined in the manuscript. My understanding is the “loop” or “coil” region of a protein, but it should be clearly defined. 6. The performance varies a lot between two test sets in Table 5. Instead of just listing the results there, I think it would be more interesting to explain these differences in greater detail. Could the authors add the LCR (%) for native sequences in this table as a reference? Nevertheless, is LCR (%) a good metric for assessing the design quality as they showed the AlphaFold2 prediction test? I also so because in CATH4.2-StructNR193, the LCR by OSCAR-design was the lowest, but its corresponding AlphaFold2 prediction test was the worst (in terms of RMSD and GDT-TS)? If we trust the global structure prediction metric, it seems that a lower LCR does not necessarily lead to a higher refolding ability. 7. In lines 693-694, “This new training set, however, led to a poorer performance for those unseen structures.” And lines 698-699 “we employed the whole training set as it improves the generalizability over the smaller training set.” Could the authors explain more reason on the poorer performance on the nonredundant, smaller training set. Is this because the “TM-score<0.4” is too strict, removing many nonredundant folds? Or is because TM-score is just a global metric and the larger set can keep more abundant local structural information which are very useful for training, and TM-score<0.4 will eliminate many useful local information? 8. The authors have tons of data, but they did not dig them very deeply, nor the Discussion section was discussed in depth. Reviewer #3: In the manuscript, the authors present a GNN-based approach, called SPIN-CGNN, to protein sequence design. The motivation of the study is to circumvent the drawbacks of constructing GNN using k-nearest neighbors with $k$ fixed beforehand. The SPIN-CGNN approach incorporates the node/edge feature settings proposed by PiFold, the message passing mechanism proposed by ProteinMPNN and GraphTrans, and two mechanisms for edge updating (second-order edge and symmetric edge) proposed by the authors themselves. The authors conducted extensive experiments (including recovery, perplexity, the predicted structures of the designed sequences, residue diversity, sequence complexity) to examine the performance of SPIN-CGNN. The idea of the SPIN-CGNN is interesting; however, I still have several concerns. Major concerns: 1. Presentation of the approach: In the manuscript, the authors described the node updating and edge updating only and did not mention the procedure for sequence generating, although they showed two blocks in Figure 1. The authors should describe the complete procedure of sequence design to make readers easy to follow. 2. Dominating factors: the approach is a mixture of a large variety of ingredients, including using virtual atoms and bond angle/length as node features, using inter-atomic vectors and distances as edge features, and using SK mechanism to merge multiple features. The mixture of multiple ingredients is useful; however, this also brings a question: Which ingredient is the dominating one that determines residue type at a position? 3. Robustness of the approach: the authors used bond angles as node features, which carry information of residue type, for example, the N-C-CA angle is within [121, 126] for histidine and within [117, 122] for leucine. This fact raises a question: is bond angle the key feature for determining residue type? If yes, does this mean that this approach relies on accurate structure by X-ray crystallography? Is this the reason that the performance of the approach is not that good on hallucinated proteins as on CATH proteins? I also suggest the authors to examine the performance of SPIN-CGNN on the structures generated by diffusion model. 4. Redundancy of node features: In addition, the authors used bond lengths as node features; however, the bond length is roughly the same for all residues. The effects of virtual atoms is also not clearly stated. Are these features really inforamtive for residue type determination? 5. Organization of the Result section: I appreciate the author’s viewpoints that, for protein sequence design, the quality of the predicted structure is much more important than sequence recovery and perplexity. The current form of the manuscript puts “impact of graph construction” and two ablation studies as the first three results. I suggest the authors to reorganize the Result section with the quality of the predicted structures (subsection 3.11) as the first subsection. I also suggest the authors to add a column into Table 1, 2, 3, 4 to show the quality the predicted structures. 6. Inconsistent figures: In both Table 2 and 3, the mean perplexity and recovery of SPIN-CGNN on CATH4.2 is 4.36 and 52.89, respectively. However, these figures are 4.05 and 54.81 in Table 4. Why? In addition, the head of Table 1, 2, 3 is “mean recovery” but the authors listed them in the form of standard variation. 7. The effect of SK: The SK technique calculates the sum of multiple features (Line 334) and then applies the multi-head mechanism on it. However, the simple addition of the features, say symmetric edge features, SOE and the extracted features (Line 281), might make one type of the features dominate the others. How about apply MLP on the concatenation of these features? 8. The effect of SOE: The message-passing mechanism aims to broadcast the information of the neighbors to the central residue even the neighbors are 1, 2, 3, … hops apart and thus naturally considers the information of secondary-order neighbors. The authors should perform case study to clearly show the effects of SOE updating. 9. Case study: I suggest the authors to provide case study to clearly state the most effective ingredients of the approach. Minor concerns: 1. Grammar errors: I suggest the authors to polish the manuscript to remove grammar errors, say: Line 405: “we performed a test to evaluated ….” 2. Position of punctuations: In Line 192, the sentence starts with a comma. Please put the comma symbol after the formula. 3. Inconsistent use of punctuations: The authors use “a)” in Line 155 but use “(a)” in Line 173. 4. “Marked as TS50” => “denoted as TS50” ********** Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No Reviewer #3: No Figure Files: While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Data Requirements: Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5. Reproducibility: To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols |
| Revision 1 |
|
Dear Dr Zhou, Thank you very much for submitting your manuscript "SPIN-CGNN: Improved fixed backbone protein design with contact map-based graph construction and contact graph neural network" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. The reviewers appreciated the attention to an important topic. Based on the reviews, your manuscript may become acceptable, providing that you modify the manuscript according to the recommendations of the third reviewer. Meanwhile, please provide appropriate references to TM-score and GDT-score which have been extensively used in the manuscript but are not necessarily known to all readers. Meanwhile, necessary references are needed for justifying some of the specific score cutoffs (e.g., TM-score>0.4). Please prepare and submit your revised manuscript within 30 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. When you are ready to resubmit, please upload the following: [1] A letter containing a detailed list of your responses to all review comments, and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out [2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file). Important additional instructions are given below your reviewer comments. Thank you again for your submission to our journal. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments. Sincerely, Yang Zhang Guest Editor PLOS Computational Biology Nir Ben-Tal Section Editor PLOS Computational Biology *********************** A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact ploscompbiol@plos.org immediately: Reviewer's Responses to Questions Comments to the Authors: Please note here if the review is uploaded as an attachment. Reviewer #1: The authors have appropriately address all my concerns and I believe the paper is acceptable for publication. Reviewer #2: The authors have tried their best to address my questions and comments. I suggest accepting the manuscript in its current form. Reviewer #3: I appreciate the authors’ efforts to address my concerns. However, I have some further concerns: 1. Dominating factors: the approach is a mixture of a large variety of ingredients and the authors’s analysis suggest that none of the ingredients is dominant and the power of the approach is essentially a “cumulative improvement from multiple changes”. This observation and conclusion are instructive and I suggest to add them to the Conclusion section. 2. Redundancy of node features: the authors used bond lengths as node features; however, the bond length is roughly the same for all residues. The effects of virtual atoms is also not clearly stated. The authors cited PiFold’s analysis to answer this question, which shows only marginal contributions by these two types of features. I suggest to add a paragraph to clearly state the effects of them. 3. Writing: The current form of the manuscript still lacks of careful proofreading. There are still many grammar errors or mistakes., e.g. — Supplementary Table S1. ….Two-layer MLP modules were applied to replace SK modules for no-SK models — Ref. 35 even has Chinese characters. — Ref. 39, 44: The major words of journal/conference names are not capitalized. ********** Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No Reviewer #3: No Figure Files: While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Data Requirements: Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5. Reproducibility: To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols References: Review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice. |
| Revision 2 |
|
Dear Dr Zhou, We are pleased to inform you that your manuscript 'SPIN-CGNN: Improved fixed backbone protein design with contact map-based graph construction and contact graph neural network' has been provisionally accepted for publication in PLOS Computational Biology. Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests. Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated. IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript. Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS. Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. Best regards, Yang Zhang Guest Editor PLOS Computational Biology Nir Ben-Tal Section Editor PLOS Computational Biology *********************************************************** |
| Formally Accepted |
|
PCOMPBIOL-D-23-01069R2 SPIN-CGNN: Improved fixed backbone protein design with contact map-based graph construction and contact graph neural network Dear Dr Zhou, I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course. The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript. Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers. Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work! With kind regards, Zsofia Freund PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .