Peer Review History

Original SubmissionMay 6, 2024
Transfer Alert

This paper was transferred from another journal. As a result, its full editorial history (including decision letters, peer reviews and author responses) may not be present.

Attachments
Attachment
Submitted filename: PUPAID_TransferRevision2_PLOS_One.docx
Decision Letter - Luca Pesce, Editor

PONE-D-24-13313PUPAID: a R + ImageJ pipeline for thorough and semi-automated processing and analysis of multi-channel immunofluorescence dataPLOS ONE

Dear Dr. Régnier,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

The revisions were positive, highlighting the manuscript's potential impact in research fields. However, while your work presents groundbreaking potential, its real advantages seem to be primarily in high-density regions. Given the clinical expertise among the authors, a detailed medical discussion would be highly beneficial.

From a reviewer's perspective, providing or at least describing the testing dataset is essential for comprehensive evaluation. Additionally, the manuscript lacks a thorough explanation of the benchmarking, which appears somewhat confusing. Improved communication on the quantitative methods used for benchmarking would significantly enhance the validation of PUPAID.

Please submit your revised manuscript by Jul 29 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.
  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.
  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Luca Pesce, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

1. When submitting your revision, we need you to address these additional requirements.

Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Thank you for uploading your study's underlying data set. Unfortunately, the repository you have noted in your Data Availability statement does not qualify as an acceptable data repository according to PLOS's standards.

At this time, please upload the minimal data set necessary to replicate your study's findings to a stable, public repository (such as figshare or Dryad) and provide us with the relevant URLs, DOIs, or accession numbers that may be used to access these data. For a list of recommended repositories and additional information on PLOS standards for data deposition, please see https://journals.plos.org/plosone/s/recommended-repositories.

3. When completing the data availability statement of the submission form, you indicated that you will make your data available on acceptance. We strongly recommend all authors decide on a data sharing plan before acceptance, as the process can be lengthy and hold up publication timelines. Please note that, though access restrictions are acceptable now, your entire data will need to be made freely accessible if your manuscript is accepted for publication. This policy applies to all data except where public deposition would breach compliance with the protocol approved by your research ethics board. If you are unable to adhere to our open data policy, please kindly revise your statement to explain your reasoning and we will seek the editor's input on an exemption. Please be assured that, once you have provided your new statement, the assessment of your exemption will not hold up the peer review process.

4. Please amend either the abstract on the online submission form (via Edit Submission) or the abstract in the manuscript so that they are identical.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Thank you for your contribution with the development of PUPAID. While your work presents groundbreaking potential, according to this version of the manuscript the real advantages appear to be primarily in high-density regions. Given the clinical expertise available among the authors, it would be highly beneficial to leverage this by providing detailed medical discussion.

From a reviewing point of view it would help to be provided with a testing dataset, or at least to describe it, for comprehensive evaluation. Additionally, the manuscript lacks a thorough explanation of the benchmarking which appears a bit confusing. I think venturing in many comparisons, different in nature, is not simple to deal with. Improved communication on the quantitative methods used for benchmarking would significantly enhance the validation of PUPAID.

I also think the paper lacks demonstration of user-friendliness, with the language used resulting in an additional barrier. Including a figure or, preferably, a video of the interface would greatly enhance understanding and usability for readers.

Lastly, the manuscript's language is quite self-referential and would benefit from a more reader-centric approach, ensuring clarity and accessibility for a broader audience. I suggest giving more scientific context and sometimes guiding the reader via bullet points.

Minor changes:

Abstract

Line 58: Would be nice for reviewers to have test files to run the code

Introduction

Line 82-83: would appreciate if the author could elaborate on what is meant by “exponential increase of the complexity”. Example: what is the role of noise and how is it evolving through time. Also it is not very clear what is an “early lack of open-source methods”

Line 88: The learning curve of coding within publicly available tools can be steep, conversely I would not define programs like Image J cumbersome to master.

Line 94: would be interesting if the author could elaborate on the relevance of Doc Parra in assessing methodology performances. Additionally, please provide further evidence of the status quo claimed here.

Lines 110-112: quite a statement about performance, please provide quantitative insights.

Dependencies

Lines 124-126: could the author stress further the decision of relying on R rather than python packages. Are R packages flexible in terms of being integrated on cloud and how suitable are they for integration with AI workflows?

Pre processing of raw data

Line 161: there are current efforts from the github community to manage such exotic formats https://github.com/cgohlke/czifile feel free to integrate in your script

Line 169: Is the “big idea” the innovation?

Line 170 and 180: It is a bit ambiguous to understand if the naming is either automatic or manual. Try to explain clearly and consider including an image solely dedicated to input data handling.

Processing of ROI

Lines 195-208: quite annoying to read through such a small amount of text calling back to so many supporting figures. Would it be possible to have a figure in main text or either organise the text differently?

Benchmarking of either PUPAID-, manual- or other state-of-the-art methods-generated

cell segmentation

Line 237: the section here seems a mix of user instructions and benchmarking. By the way, consider including a user interaction section along with UI images, would probably be very beneficial to promote the adoption of the tool.

Line 247: provide quantitative details about the test datasets

Line 252-253: what is the error on the single metric you are using to claim significance? There are hints in the images but this information is a bit tough to retrieve on the spot.

Results

Line 319: how is the past application of a novel tool a result? Consider providing a clear explanation. You could refer to validation on published studies, and simply include the reference

Line 342: Again this sounds like a validation rather than result section.

Reviewer #2: The manuscript PONE-D-24-13313 by P. Régnier and colleagues presents PUPAID, a workflow developed in R and ImageJ for semi-automated processing and analysis of multi-channel immunofluorescence data. They explain its workflow, which is user-friendly by incorporating an optional R Shiny-based interactive application for those not proficient in R. PUPAID performance is validated on few datasets and compared with state-of-the-art methods like StarDist or Cellpose, identifying conditions where it outperforms them, especially in dense areas. It exports single-cell data as FCS files, compatible with various cytometry analysis software. PUPAID is available as a GPLv3-licensed R package on GitHub. The work is well presented and with convincing experimental results and it is well written. The installation and usage is well documented and reproducible.

I support the publication of the manuscript without additional changes.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Vladislav Gavryusev

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachments
Attachment
Submitted filename: Comment to authors PLOSone.docx
Revision 1

Response to Editor and Reviewers

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

The revisions were positive, highlighting the manuscript's potential impact in research fields. However, while your work presents groundbreaking potential, its real advantages seem to be primarily in high-density regions. Given the clinical expertise among the authors, a detailed medical discussion would be highly beneficial.

From a reviewer's perspective, providing or at least describing the testing dataset is essential for comprehensive evaluation. Additionally, the manuscript lacks a thorough explanation of the benchmarking, which appears somewhat confusing. Improved communication on the quantitative methods used for benchmarking would significantly enhance the validation of PUPAID.

Journal Requirements:

1. When submitting your revision, we need you to address these additional requirements.

Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

As requested by the Editor, we profoundly edited our manuscript to answer to the reviewers’ comments as well as to fully comply with the PLOS One journal requirements regarding the formatting of the manuscript. The revised version of the manuscript should fully embrace the journal requirements based on the two PDF resources the Editor provided to us.

2. Thank you for uploading your study's underlying data set. Unfortunately, the repository you have noted in your Data Availability statement does not qualify as an acceptable data repository according to PLOS's standards.

At this time, please upload the minimal data set necessary to replicate your study's findings to a stable, public repository (such as figshare or Dryad) and provide us with the relevant URLs, DOIs, or accession numbers that may be used to access these data. For a list of recommended repositories and additional information on PLOS standards for data deposition, please see https://journals.plos.org/plosone/s/recommended-repositories.

As requested by the Editor, and given the provided URL regarding the recommended repositories, we uploaded the example dataset on the Data Station Life Sciences repository hosted by the Data Archiving and Networked Services (DANS). The associated accession URL (https://doi.org/10.17026/LS/7XQFAT) was indeed added within the revised version of the manuscript.

3. When completing the data availability statement of the submission form, you indicated that you will make your data available on acceptance. We strongly recommend all authors decide on a data sharing plan before acceptance, as the process can be lengthy and hold up publication timelines. Please note that, though access restrictions are acceptable now, your entire data will need to be made freely accessible if your manuscript is accepted for publication. This policy applies to all data except where public deposition would breach compliance with the protocol approved by your research ethics board. If you are unable to adhere to our open data policy, please kindly revise your statement to explain your reasoning and we will seek the editor's input on an exemption. Please be assured that, once you have provided your new statement, the assessment of your exemption will not hold up the peer review process.

We thank the Editor for his remark. As requested in the previous point, we already uploaded the full example dataset on the Data Station Life Sciences repository hosted by the Data Archiving and Networked Services (DANS), which will be accessible via the following URL: https://doi.org/10.17026/LS/7XQFAT. Regarding the GitHub repository, all authors confirm that the code with be fully available if the manuscript is accepted for publication. Concretely, the developer and maintainer of PUPAID (Dr. Paul Régnier actually) only has to toggle a setting from “Private” to “Public” in his GitHub repository profile, which will instantly make the full code available to everyone. He again wanted to reassure the Editor that he will do this task without delay if the manuscript is accepted for publication.

4. Please amend either the abstract on the online submission form (via Edit Submission) or the abstract in the manuscript so that they are identical.

We wanted to apologize for this point, as this was not intentional at all. The corresponding author probably copied and pasted within the online submission form a version of the abstract which was not the very last one. We carefully checked both abstracts and made sure they were identical between the manuscript and the online submission system.

Reviewers’s comments to the Author

Reviewer #1:

Thank you for your contribution with the development of PUPAID. While your work presents groundbreaking potential, according to this version of the manuscript the real advantages appear to be primarily in high-density regions. Given the clinical expertise available among the authors, it would be highly beneficial to leverage this by providing detailed medical discussion.

We thank the reviewer for his very important remark. To discuss this interesting point, we added a new Potential impact of PUPAID on the clinical and medical fields section in the manuscript which details how PUPAID could be useful in the clinical and medical fields. We notably expose here tissues and their associated diseases for which immunofluorescence is widely used for both diagnostic and research purposes, and for which standardized and robust analysis methods could potentially help a lot to better understand the diseases’ pathophysiology, but also to accelerate the discovery of new biomarkers and further refine the classification of patients.

From a reviewing point of view it would help to be provided with a testing dataset, or at least to describe it, for comprehensive evaluation. Additionally, the manuscript lacks a thorough explanation of the benchmarking which appears a bit confusing. I think venturing in many comparisons, different in nature, is not simple to deal with. Improved communication on the quantitative methods used for benchmarking would significantly enhance the validation of PUPAID.

We thank the reviewer for his remark and made modifications to improve the points he highlighted. First, to ease readers understanding, we transferred and renamed the Test dataset section present in the Supplemental Material and Methods to a new Example dataset for learning and test purposes subsection in the main manuscript in the Material and methods section in order to present the example dataset, and notably what kind of organ/patient is the tissue from, how it was stained (technology, antibodies and Opal reagents) and how it was acquired and processed (instrument, software and demultiplexing step). We also indicated the new URL from which users can download the dataset.

We additionally moved every supplemental methods from the Supplemental Material and Methods to the main manuscript. This lengthens the manuscript a bit but allows to avoid switching between files for an explanation of a precise point. In a similar manner, we decided to integrate the old Figure S1 as a new Figure 2 in the main manuscript, to directly provide to readers the totality of the benchmarking data.

Second, we were unsure about what the reviewer specifically referred to when writing about the “benchmarking”, because the manuscript details two different ones. The first benchmarking was related to the CLAHE algorithm and helped to demonstrate that this step was pertinent and enhanced the subsequent cell segmentation through PUPAID. The second benchmarking was about the performance of PUPAID (with different σ ratios) as compared to manual cell segmentation but also to other well-established methods: StarDist and Cellpose. To improve the overall readability of the manuscript, we decided to edit both parts in the revised version of the manuscript, even if one was not directly targeted by the reviewer. In a sense, his remark tells us that the manuscript was probably not organized properly. To summarize, we transferred the Comparison of masks generated with different cell segmentation methods section originally present in the Supplemental Material and Methods to the main manuscript, in the Benchmarking of the different cell segmentation approaches subsection of the Material and methods section. We also extracted the CLAHE algorithm benchmarking part to a dedicated subsection between the Processing of ROI and Benchmarking of either PUPAID-, manual- or other state-of-the-art methods-generated cell segmentation subsections. This way, the methodology used for both benchmarking is totally visible and independent of the other sections, and the CLAHE-related section comes after the full description of the ROI processing. We believe this will improve the readability of the manuscript. These modifications indeed led to integrating the old Figure S1 as the new Figure 2, whereas the old Figure 2 is now the new Figure 3 and the old Figure 3 is now the new Figure 4.

I also think the paper lacks demonstration of user-friendliness, with the language used resulting in an additional barrier. Including a figure or, preferably, a video of the interface would greatly enhance understanding and usability for readers.

We greatly thank the reviewer for his very pertinent remark about the user-friendliness of PUPAID. Even if it was not as easy as one could think, we finally managed to generate a video of roughly 14 minutes which shows in live the full workflow of PUPAID applied on the example dataset. This video is accessible from the following YouTube link: https://youtu.be/58Tm54OVP-g.

Lastly, the manuscript's language is quite self-referential and would benefit from a more reader-centric approach, ensuring clarity and accessibility for a broader audience. I suggest giving more scientific context and sometimes guiding the reader via bullet points.

We greatly thank the reviewer for his constructive remark. To make our manuscript easier to read, we deeply modified its structure and content, as previously stated. First, the (now) full compliance with the PLOS One journal format helped a lot to restructure our work and our thoughts. Second, we rewrote entire sections of the manuscript to give more details (or less in some paragraphs) and to make a lot of sentences clearer. We also moved/switched full sentences/paragraphs to other (or even new) subsections to better structure the manuscript. Third, to enhance clarity and readability, we also put as main figures and methods the old supplemental materials originally submitted. This will ensure that readers will not have to permanently switch between documents when they read our manuscript. Fourth, we modified as much as we could the figures to make graphs and texts as big as possible to ease the reading. Fifth, we implemented a new subsection to add more scientific context regarding PUPAID’s utility and impact, notably in the medical field.

Minor changes:

Abstract

Line 58: Would be nice for reviewers to have test files to run the code

We thank the reviewer for his remark but wanted to clarify that test files (in this case, TIFF-formatted images containing multi-channel immunofluorescence data) were already available during the last submission of this manuscript, as mentioned in the last paragraph of the old Availability and future directions section of the originally submitted manuscript. If the reviewer encountered difficulty to access the files using the provided URL within the last weeks/months, we wanted to apologize for this issue, which was indeed not intentional. To address this problem in the long term, and in accordance with the Editor, we uploaded the test dataset to the Data Station Life Sciences repository hosted by the Data Archiving and Networked Services (DANS), where it should be available at all times. The test dataset will be totally accessible via the following URL: https://doi.org/10.17026/LS/7XQFAT.

Introduction

Line 82-83: would appreciate if the author could elaborate on what is meant by “exponential increase of the complexity”. Example: what is the role of noise and how is it evolving through time. Also it is not very clear what is an “early lack of open-source methods”

To be more precise, we wanted to underline the fact that over time, and especially during the two last decades, multiplex immunofluorescence underwent massive improvements together with a great diversification of methods, thus leading to the increase of the number of simultaneously acquired markers as well as the overall resolution and number of co-analyzed samples. We rephrased this sentence to be clearer and added references to emphasize on this point. We also removed the early lack of open-source methods part, because if it was true 10 to 20 years ago, it does not seem to be totally valid anymore, especially when we look at the plethora of analysis methods that were recently developed for this precise kind of purpose.

Line 88: The learning curve of coding within publicly available tools can be steep, conversely I would not define programs like Image J cumbersome to master.

We wanted to clarify what is said in this part of the manuscript. We did not mean to say that Image J is difficult to master. Actually, this software is very well programmed and organized, and offers the ability to quickly analyze heavy images with a broad variety of tools. However, if one needs to analyze dozens or hundreds of images using the same methodology in Image J, the need for automatization through macros and scripts (thus necessarily involving programming) is almost mandatory. To make this point of view clearer, we also rephrased this part.

Line 94: would be interesting if the author could elaborate on the relevance of Doc Parra in assessing methodology performances. Additionally, please provide further evidence of the status quo claimed here.

We thank the reviewer for his very constructive comment. Our initial sentence was meant to highlight the fact that the plethora of available analysis methods for immunofluorescence assays could potentially lead to the biased selection by researchers/scientists of methods that actually “produce” convenient results and thus avoid the “least performant” ones. We wanted to underline the notion that workflows which process data from A to Z (pre-processing, processing, normalization, analysis, results export, etc.) are one way to intent standardization of such analyses and avoid potential issues and biases. We thought this was relevant to write at first, but all your remarks taken as a whole finally lead us to doubt about the true relevance of this sentence in the introduction. In essence, our message is slightly suggested in the sentence the vast majority of the available methods only seems to propose a given kind of analysis and almost never include actual pre-processing and processing of the raw images written right before. In consequence, we thought it was better to remove the sentence related to Dr. Parra and thus rewrote the two last paragraphs of the introduction to make it clearer.

Lines 110-112: quite a statement about performance, please provide quantitative insights.

We thank the reviewer for his remark and added some quantitative insights in the last paragraph of the introduction but also added a new Table 2 which summarizes the efficacy of StarDist, Cellpose and PUPAID (using its most balanced 1:1.4 σ ratio and an IoU threshold of 0.4) independently for each measurement (% of detected objects and each of the four mentioned metrics) and in average. This will help readers to have a simpler “take home” message regarding PUPAID’s efficacy as compared to other well-established cell segmentation methods.

Dependencies

Lines 124-126:

Attachments
Attachment
Submitted filename: PUPAID_PLOS_One_Revision.docx
Decision Letter - Luca Pesce, Editor

PUPAID: a R + ImageJ pipeline for thorough and semi-automated processing and analysis of multi-channel immunofluorescence data

PONE-D-24-13313R1

Dear Dr. Régnier,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Luca Pesce, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Great job, really appreciated the hard work on making the manuscript more readable and sharing data from a more insightful perspective.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

**********

Formally Accepted
Acceptance Letter - Luca Pesce, Editor

PONE-D-24-13313R1

PLOS ONE

Dear Dr. Régnier,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Luca Pesce

Academic Editor

PLOS ONE

Open letter on the publication of peer review reports

PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.

We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.

Learn more at ASAPbio .