Peer Review History

Original SubmissionMarch 24, 2024
Decision Letter - Achraf El Allali, Editor

PONE-D-24-11846Comparisons of performances of Structural Variants Detection Algorithms in Solitary or Combination StrategyPLOS ONE

Dear Dr. Duan,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please address the reviewers' comments carefully as they have reaised some serious questions. Please explain why the need for another SV calling benchmark using existing datasets. I also noticed that you have not included any machine learning based algorithms. 

Please submit your revised manuscript by Jul 12 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.
  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.
  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Achraf El Allali, PhD

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match. 

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section.

3. In the online submission form, you indicated that "For any further requests and questions, the data used and/or analyzed during the current study are available from the corresponding author upon reasonable request."

All PLOS journals now require all data underlying the findings described in their manuscript to be freely available to other researchers, either 1. In a public repository, 2. Within the manuscript itself, or 3. Uploaded as supplementary information.

This policy applies to all data except where public deposition would breach compliance with the protocol approved by your research ethics board. If your data cannot be made publicly available for ethical or legal reasons (e.g., public availability would compromise patient privacy), please explain your reasons on resubmission and your exemption request will be escalated for approval.

4. When completing the data availability statement of the submission form, you indicated that you will make your data available on acceptance. We strongly recommend all authors decide on a data sharing plan before acceptance, as the process can be lengthy and hold up publication timelines. Please note that, though access restrictions are acceptable now, your entire data will need to be made freely accessible if your manuscript is accepted for publication. This policy applies to all data except where public deposition would breach compliance with the protocol approved by your research ethics board. If you are unable to adhere to our open data policy, please kindly revise your statement to explain your reasoning and we will seek the editor's input on an exemption. Please be assured that, once you have provided your new statement, the assessment of your exemption will not hold up the peer review process.

5. Please include your full ethics statement in the ‘Methods’ section of your manuscript file. In your statement, please include the full name of the IRB or ethics committee who approved or waived your study, as well as whether or not you obtained informed written or verbal consent. If consent was waived for your study, please include this information in your statement as well.

6. We notice that your supplementary tables are included in the manuscript file. Please remove them and upload them with the file type 'Supporting Information'. Please ensure that each Supporting Information file has a legend listed in the manuscript after the references list.

Additional Editor Comments:

Please address the reviewer's comments carefully. Please explain why the need for another benchmarking tests. I also noticed that you have not included any machine learning based algorithms.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: No

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The manuscript describes a benchmark experiment to detect structural variants from short reads. It uses the HG002 benchmark dataset of the Genome in a bottle project as gold standard. Due to lack of information in the gold standard or lack of events in some of the evaluated tools, this benchmark only takes into account deletions between 50 and 10kbp.

Although I understand that reviews for PloS One should not be emphasized in novelty, I think benchmarking of SVs with GIAB data is one extreme case. I do not see a point on publishing over and over the same benchmark experiment with the exact same gold standard. Just by a simple google search with the words “benchmark structural variants detection”, I obtained the following papers:

- Zook et al., 2020. https://doi.org/10.1038/s41587-020-0538-8

- Sarwal et al., 2022. https://doi.org/10.1093/bib/bbac492

- Fang et al., 2024. https://doi.org/10.1186/s12967-024-04865-w

- Cameron et al., 2019. https://doi.org/10.1038/s41467-019-11146-4

The two more recent papers are not even cited in the literature. The main issue of making all benchmarking experiments on only one dataset as gold standard is that the development of the different tools is likely to be biased towards that particular dataset, which is likely to produce overfitting. Moreover, it is at least strange that even having an almost perfect diploid assembly for this dataset, there are still no gold-standard calls for inversions and duplications. I think that the authors should come up with at least one benchmark experiment different than GIAB, to really assess the accuracy of the different tools in cases different than those probably used to tune these tools. One source to at least improve the genetic diversity of the samples is to perform benchmarking using the datasets generated by the HGSVC2 consortium (see https://doi.org/10.1038/s41467-018-08148-z and https://doi.org/10.1126/science.abf7117)

I have some additional comments, as follows.

Lines 43 – 45, and 466 - 469. It is not clear which is the evidence to support the statements made in these lines of the abstract and the conclusions. They seem to be based solely on the number of variants produced by each tool, before the assessment using the gold-standard. In particular it is impossible to draw conclusions about translocations without a gold standard to assess the accuracy of these events.

Lines 52-56: Taking into account that the decisions regarding variant detection and interpretation have been rather conservative, I really doubt that a method in which the false positives of all tools are retained would be the recommended method for diagnosis in clinical settings.

Lines 94 – 96. This statement is less true each day. Both PacBio and Nanopore have made important cost reductions and improvements in quality, reducing the cost difference with Illumina. Taking into account that SV calling with long reads has a clearly superior accuracy, compared to SV calling with short reads, I would at least discuss if it is really worth to keep considering Illumina for SV detection in clinical settings.

Lines 99 – 104. Although the issue with annotation of SVs is true, the paper does not address this issue. I recommend removing these lines, and instead writing a reduced version of what is currently on the discussion, providing a brief description of the algorithms implemented in the SV calling tools.

Lines 157 – 168: The procedures to merge results of the different tools need more clarity. What happen if for two events separated 1Kbp there is one intermediate event exactly in the middle?. It is not clear which percentage of (reciprocal) intersection is required to call two events the same. It is also not clear how the 500bp threshold was chosen. The merging called intersection is not really an intersection of the methods. It is more like a two out of n kind of union. It should not be called intersection in part because it does not make sense that the intersection of five methods resulted in more events than the intersection between three methods.

Lines 180 – 184. The procedure to call a variant true positive has the same issues as the procedure to merge calls. The 500bp limit looks arbitrary, especially if no reciprocal overlap is required. Moreover it looks too relaxed for relatively small events (lets say less than 150 bp), which are a large percentage of the total events.

Lines 271 – 274. It does not make sense to restrict the benchmark only to deletions just because two methods do not call insertions. Benchmarking of insertions should be included for the methods that predict insertions. Because comparing true and predicted insertions is more challenging, compared to deletions, it is important to define and present clearly specific rules to perform both merging and intersection of insertions.

Lines 313- 316. I would not say that the unsatisfactory results are compared to Dragen. I would say that the results are unsatisfactory overall.

Line 337. It is not surprising that the maximum is 500 if that correspond to the threshold set by the method. Again, it looks too relaxed for relatively small variants.

Lines 461 – 463. I do not see evidence of anything different or special in this work to say that it was focused on clinical applicability. There is no information on a disease or condition that could be related to the SVs detected on HG002. The authors should clarify what is their meaning of “clinical applicability”.

Finally, most of the current discussion could be written without the results of this study. Most of the text could be moved and summarized in the introduction. The text includes many sentences with qualitative assessments of each tool that are not related to the results, and in some cases are even contradictory. The description of the tools could be summarized in a couple of paragraphs of the introduction, avoiding a-priori qualitative assessments about the performance of the algorithms. Once the other comments are solved, especially finding additional gold standards and evaluating more types of variants, an actual discussion could be written based on the results of the new experiments.

Minor comments

- Figures and tables: The distribution between main and supplementary materials looks arbitrary. If supplementary tables are going to be embedded in the text, then they should not be called supplementary.

- Lines 150 -151. The sentence itself does not say much. I do not understand what is stringent about the filtering procedure. The filters mentioned in the text look fairly standard.

- Figures 1 and 2 should be composed to understand why the subfigures should go together. The figures look disorganized in the submitted version.

- Font types and resolution should be improved for Figures 3A and 3B

- The supplementary figure 3 has essentially the same information as Figure 2A. Moreover, it does not make sense to include bars for inversions and duplications if the dataset does not include these events.

Reviewer #2: Duan et al., assessed six SV calling algorithms, including DRAGEN, DELLY, LUMPY, DRIDSS, SvABA, and Manta. The assessment was done on HG002/NA24385 sample from Genome in a Bottle (GIAB).

Major Comments:

1. Reference Genome: The assessment was done on alignments against GRCh37, which is less complete than GRCh38. Please explain why GRCh38 was not used.

2. SV Truth Set: GIAB should have an officially released SV list. Why was this not used as the Truth set instead of defining your own?

3. Mendelian Violation Experiment: Since the sample used for the benchmark is a son from a trio, performing a Mendelian violation experiment by comparing with the parents would add value.

4. Conclusion Alignment: The conclusion does not fully align with the title "Comprehensive evaluation and characterisation of short read general-purpose structural variant calling software."

5. Translocation Assessment: Why were translocations not included in the assessment?

6. SvABA Reporting: Does SvABA report SVs in BND format? If so, how did you convert BND to other SV types?

7. Duplicate Records: Lines 147-148 mention that duplicated records will be removed. Under what conditions would a calling algorithm report duplicated records?

8. Duplication and Insertion Reporting: Calling algorithms may report duplications as insertions. How do you handle this case if you only combine SVs of the same type?

9. Size Range Limitation: Please explain why the comparison was limited to the range of 50 to 10k base pairs.

Minor Comments:

1. To be precise, replace “Jewish” with “Ashkenazim.”

2. The authors mention that long-read sequencing in clinical diagnostics is hindered by cost constraints, implying this study would cover this application. However, there is no coverage of this aspect in the manuscript.

3. Please spell out SNV and indel the first time they are used. Regarding SV, once it is defined, there is no need to define it every time it appears.

4. Clarify what you mean by “high-throughput analysis” in the third paragraph of the Background section (Line 78).

5. Replace “paired short reads” with “paired-end short reads” (Line 139).

6. Ensure consistent use of "nucleotide" or "base pair."

7. The term “least POS” is unclear (Line 161).

8. The statement “We systematically searched scientific articles in PubMed” is ambiguous. Consider removing it.

9. The Input and Output columns are identical for all callers, making them uninformative. Consider removing these columns.

10. Table S3 is redundant as its information is covered by Table S2.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Wan-Ping Lee

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Revision 1

We greatly thank the editor and reviewers for the thoughtful comments of our manuscript. Following with all your kind suggestions, we have made improvement to our manuscript. The detailed point-by-point response to the comments are listed on the following pages. All the changes in revised manuscript are indicated by red color, and we also provide a clear version of our manuscript in the submission package. We really hope the revision will make the paper be more suitable for publication in PLOS ONE and look forward to hear from you in the near future. Your kind consideration will be highly appreciated!

Attachments
Attachment
Submitted filename: point-by-point response letter to Plos One.docx
Decision Letter - Achraf El Allali, Editor

PONE-D-24-11846R1Comparisons of performances of structural variants detection algorithms in solitary or combination strategyPLOS ONE

Dear Dr. Chen,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Oct 12 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.
  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.
  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Achraf El Allali, PhD

Academic Editor

PLOS ONE

Additional Editor Comments:

Please address the reviewers comments.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #3: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The manuscript describes a benchmark experiment comparing tools to perform detection of SVs from short read data. The authors performed important changes in the manuscript to address most of my previous concerns, especially those related to the use of HG002 as the only benchmark dataset. They also fixed most empty, vague and even incorrect phrases likely produced by the use of large language models for automated writing (I recommend avoiding the use of such models). Regarding novelty, I would assume that the manuscript is interesting for the journal editors, taking into account that it was send again for another round of review.

I have the following remaining comments for improvement of the study, the last two related to the support of some of the statements included now in the discussion:

1. I do not see a point on removing variants longer than 10kbp from the analysis. The number of gold standard variants is small, and hence it will not affect importantly the calculations of recall. However, figure 2 suggest that there is a relatively large number of false positive variants in that size range. These false variants are important (especially if they are deletions) because they can be interpreted as important losses of DNA (and possibly genes) that could be falsely related to a trait. It would be great if the authors show an example of at least one of these predictions to rationalize why the software tools make mistakes in this size range.

2. I do not understand why variants with lengths smaller than 50bp are filtered out for all tools except GRIDSS. This makes unnecessarily complicated the interpretation of the number of variants produced by each tool. In figure 2, Figure 2A will look much better if small variants reported by GRIDSS are filtered out. In figure 2B it is not clear why different scales are used in the Y axis for the different tools if all tools report numbers within the same orders of magnitude (once the results of GRIDSS are fixed).

3. Referring to detection of SVs from long reads, the authors mention in the discussion that “A more comprehensive assessment of these technologies is still lacking”. First of all it is not clear compared to to what refers the term “more comprehensive”. More importantly, the statement is arguably false. All papers presenting new tools for SV detection from long reads include benchmark experiments (see dois 10.1093/gigascience/giad112, 10.1186/s13059-020-02107-y, and 10.1038/s41592-018-0001-7 for example ). Comparing to the accuracies presented in this work on the same benchmark datasets, it is clear that SV detection from long reads is much more accurate than SV detection from short reads, regardless the tool used for analysis. This should be clearly mentioned in the discussion, referring to the appropriate papers.

4. The authors mention in the discussion that the analysis of long reads takes a longer runtime compared to the analysis of short reads, referring to bwa. This is likely a wrong statement. First of all, it is not clear from the text why bwa is mentioned, taking into account that bwa would be the mapping tool and hence it should be compared to tools such as minimap. More importantly, the authors do not provide any result on computational efficiency of the different methods. Since this is actually an important outcome of every benchmark experiment, the authors should include a comparison of computational efficiency among the methods considered in this study. It is understandable that the runtime of their merging strategies will be at least the sum of the runtimes of the individual variant callers. If the authors want to make a statement comparing computational efficiency between the analysis of short and long reads, then they should provide comparisons including software to detect SVs from long reads.

I have some additional minor comments (line numbers refer to the version with control change):

Pages 15, 17 and 20: I think it is not worth to spell every single value of true positives, false positives and false negatives in the text, especially if you also have tables in the main text with the values. You could use ranges and only report the most important numbers that you want to highlight.

Line 24: Why the role of SVs is “pivotal”

Line 50: Base pairs is written twice

Line 82. A noun is missing before “… become more accessible”

Line 94: It is not clear which is the “sector” mentioned in the sentence.

Line 99: A simple sentence should not start with “Although”.

Line 105. The a-priori qualitative statement about SvABA probably needs a reference, especially taking into account that later SvABA turns out to be one of the worst tool in the comparison.

Line 147 - 149: The first sentence is written in active voice and the second sentence is written in passive voice. Either one can be use but do not mix up.

Line 161 - 163: The sentence is confusing. It may be better to explain first the issue and then the solution.

Line 201. strategies is misspelled

Line 246. Please explain in the methods how “state-of-the art” is defined and evaluated as a software quality attribute to select software tools.

Line 304. Please provide an statistical test and p-value to assess why the number of short variants reported by GRIDSS is “significant”. Please explain compared to what the number is significantly higher or smaller.

Lines 340 - 342: Given that the gold standard is a fixed dataset, the trend of false negatives can be directly inferred from the trend of true positives.

Line 378. The first sentence seems to need citing a reference

Line 387. It is worth to actually cite the GIAB study in this sentence

Line 406. “Close” is misspelled

Line 432 – 434. Again , given that the gold standard is a fixed dataset, the trend of false negatives can be directly inferred from the trend of true positives.

Line 451: Worst than what?

Line 478: The paper actually compares the insertions from Manta with those reported by DRAGEN. Please rephrase.

Line 490: Why the similarities are “striking”

Line 492: Significantly more intricate than what?. Please provide an statistical test and p-value to assess significance.

Line 533. The sentence is vague. Why the union method has “comprehensive coverage”?. Moreover, the sentence is contradictory with the results and even with the previous sentence. The previous sentence says that the performance of individual callers is “unsatisfactory due to a poor recall rate”, and then this sentence says that the individual callers have “strong performance”.

Reviewer #3: The author evaluated six classic SV detection tools on second-generation sequencing data and explored the impact of merging strategies on improving SV detection performance. Below are some suggestions regarding this work:

1. The author's combination experiment results indicate that the combination strategies indeed improved performance by increasing either recall or precision. However, while this conclusion is evident, the results of the two merging strategies still did not surpass the performance of the single tool, DRAGEN. To further enhance SV detection performance, the author should explore the feasibility of combination strategies based on DRAGEN’s SV detection results.

2. According to the methods described, the author’s validation approach for the combination strategy involves retaining the merged SVs from each VCF file and then calculating the overall TP, FP, and FN values. A more direct approach would be to merge the results into a single VCF file and then calculate the F1 score. The author might consider comparing the outcomes of these two validation methods.

3. In the results section, the extensive listing of data might obscure the main conclusions. It is recommended that the author focus on summarizing the key conclusions, present comparative data in tables, and highlight the best results in bold.

4. The discussion section devotes significant space to background and motivation, which would be better emphasized in the introduction. It is advisable to reposition this content accordingly.

5. The author is advised to refine the abstract for clarity and conciseness.

6. The low recall rate of DELLY and GRIDSS tools in detecting INS variants is unconvincing. The author should verify the data or provide a reasonable explanation based on the literature.

7. All formulas in the manuscript should be numbered for reference and citation purposes.

8. The term ‘INFOa’ on page 8 of the manuscript should be corrected to ‘INFO’.

9. The representation of recall, precision, and F1 scores should be consistent. For example, if the F1 score is 0.67, then the recall should be represented as 0.63, not as a percentage like 62.57.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Revision 2

We greatly thank the editor and reviewers for the thoughtful comments of our manuscript. Following with all your kind suggestions, we have made improvement to our manuscript. The detailed point-by-point response to the comments are listed on the following pages. All the changes in revised manuscript are indicated by red color, and we also provide a clear version of our manuscript in the submission package. We really hope the revision will make the paper be more suitable for publication in PLOS ONE and look forward to hear from you in the near future. Your kind consideration will be highly appreciated!

Attachments
Attachment
Submitted filename: point-by-point response letter to Plos One.docx
Decision Letter - Achraf El Allali, Editor

PONE-D-24-11846R2Comparisons of performances of structural variants detection algorithms in solitary or combination strategyPLOS ONE

Dear Dr. Chen,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Dec 06 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.
  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.
  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Achraf El Allali, PhD

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments:

The reviewers have finished their review and have minor comments that I include here.

R1

The authors address properly my previous comments. As a final note, please make sure that the final versions of the figures have good quality and try to increase font sizes as much as possible. In particular the numbers in figure 4 are very difficult to see.

R2

1. Please conduct a thorough review of the format and text throughout the document. Specifically, the citation at line 75 should be placed before the period, and the comma usage at line 274 needs to be corrected.

2. I recommend reordering the references to align as closely as possible with the order in which they appear in the text.

3. The manuscript primarily discusses tools published before 2020; however, significant advancements have been made in SV detection tools based on second-generation sequencing data in recent years. For instance, tools like "PanSVR: Pan-genome augmented short read realignment for sensitive detection of structural variations" and "Psi-Caller: a lightweight short read-based variant caller with high speed and accuracy" represent the latest research outcomes and should be discussed more thoroughly in the text. Additionally, the discussion on third-generation sequencing data is insufficient. It is advisable to incorporate discussions on the latest tools for SV detection using third-generation sequencing data, such as "SVDF: enhancing structural variation detection from long-read sequencing via automatic filtering strategies" and "Kled: an ultra-fast and sensitive structural variant detection tool for long-read sequencing data," to enrich and enhance the discussion section.

In the spirit of what reviewer 2 have mentioned, there are some other tools that use machine learning and post processing approaches such as the following doi: 10.4172/0974-276X.1000488 and https://doi.org/10.1177/117793221989295

Since Plos one does not have a proofreading step, please check all the grammar and make sure all figures have the best quality possible.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors address properly my previous comments. As a final note, please make sure that the final versions of the figures have good quality and try to increase font sizes as much as possible. In particular the numbers in figure 4 are very difficult to see.

Reviewer #3: The author has made revisions in response to the issues I raised, however, I still I have the following remaining comments:

1. Please conduct a thorough review of the format and text throughout the document. Specifically, the citation at line 75 should be placed before the period, and the comma usage at line 274 needs to be corrected.

2. I recommend reordering the references to align as closely as possible with the order in which they appear in the text.

3. The manuscript primarily discusses tools published before 2020; however, significant advancements have been made in SV detection tools based on second-generation sequencing data in recent years. For instance, tools like "PanSVR: Pan-genome augmented short read realignment for sensitive detection of structural variations" and "Psi-Caller: a lightweight short read-based variant caller with high speed and accuracy" represent the latest research outcomes and should be discussed more thoroughly in the text. Additionally, the discussion on third-generation sequencing data is insufficient. It is advisable to incorporate discussions on the latest tools for SV detection using third-generation sequencing data, such as "SVDF: enhancing structural variation detection from long-read sequencing via automatic filtering strategies" and "Kled: an ultra-fast and sensitive structural variant detection tool for long-read sequencing data," to enrich and enhance the discussion section.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Jorge Duitama

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Revision 3

We greatly thank the editor and reviewers for the thoughtful comments of our manuscript. Following with all your kind suggestions, we have made improvement to our manuscript. The detailed point-by-point response to the comments are listed on the following pages. All the changes in revised manuscript are indicated by red color, and we also provide a clear version of our manuscript in the submission package. We really hope the revision will make the paper be more suitable for publication in PLOS ONE and look forward to hear from you in the near future. Your kind consideration will be highly appreciated!

Attachments
Attachment
Submitted filename: Response to Reviewers.docx
Decision Letter - Achraf El Allali, Editor

Comparisons of performances of structural variants detection algorithms in solitary or combination strategy

PONE-D-24-11846R3

Dear Dr. Chen,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Achraf El Allali, PhD

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Line 472, "may could" should be replaced by one of the two words

Line 478 you should not use 10^ but rather use the superscript in word or latex.

Reviewers' comments:

Formally Accepted
Acceptance Letter - Achraf El Allali, Editor

PONE-D-24-11846R3

PLOS ONE

Dear Dr. Chen,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Achraf El Allali

Academic Editor

PLOS ONE

Open letter on the publication of peer review reports

PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.

We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.

Learn more at ASAPbio .