Peer Review History

Original SubmissionSeptember 10, 2025
Decision Letter - Anna Bernasconi, Editor

PONE-D-25-49086 Using GDACS to Anticipate Clinical and Operational Burden after Earthquakes: A Global Event-Level Analysis (2020–2024) PLOS ONE

Dear Dr. Aykut,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jan 09 2026 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.
  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.
  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Anna Bernasconi, PhD

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that funding information should not appear in any section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. Please remove any funding-related text from the manuscript.

 3. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

4. If the reviewer comments include a recommendation to cite specific previously published works, please review and evaluate these publications to determine whether they are relevant and should be cited. There is no requirement to cite these works unless the editor has indicated otherwise. 

5. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments:

Dear authors, we have now received two reports. Please address all the points raised by Rev. 2 to strengthen the message of your contribution. Along with a point-by-point response, please attach a revised version of your manuscript with changes clearly marked in a different color. We will be ready to reassess the work.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The article is well written and provides useful information. the data provided and conclusions drawn will be quite useful for further work. It can be a good study to study and work on aftermaths of a natural disaster

Reviewer #2: Thank you for submitting this valuable manuscript. The integration of GDACS disaster alerts with health-service and wastewater operational data represents a meaningful contribution. Several areas, however, would benefit from clarification or refinement to improve rigor and reproducibility.

1. GDACS Indicator Selection and Operationalization

The manuscript mentions GDACS alerts and “subnational exposure” but does not clearly describe:

which GDACS parameters were used (e.g., alert score, exposure index, population affected),

how thresholds were defined,

and whether the chosen thresholds have prior validation.

Please add a concise table summarizing:

GDACS variables included,

units,

cut-offs,

rationale for selection.

This will increase transparency and facilitate replication.

2. Temporal Alignment Between GDACS Signals and Outcomes

The methods should clarify the exact temporal resolution used when linking GDACS alerts to:

clinical service indicators,

wastewater operational disruptions.

It is unclear whether you used:

same-day correspondence,

lagged effects (e.g., 1–3 days),

cumulative exposure windows.

Clarifying the temporal structure and providing justification (e.g., sensitivity analyses) will strengthen the analytical foundation.

3. Definition of Wastewater “Operational Impacts”

The outcomes described as “operational wastewater impacts” appear broad.

Please specify:

whether these represent flow interruptions, contamination events, pump failure, overflow, staffing disruptions, or other categories,

how each outcome was coded,

and whether events of different severity were analyzed separately or pooled.

A short operational glossary would be very helpful.

4. Statistical Modelling and Uncertainty Reporting

The statistical section briefly mentions associations but does not detail:

model type (e.g., logistic regression, GLM, Poisson, time-series),

covariates included,

handling of missing data,

validation approach.

Consider explicitly reporting:

effect estimates with 95% confidence intervals,

model diagnostics,

and sensitivity checks (e.g., excluding borderline GDACS events).

These additions would substantially improve interpretability.

5. Data Sources and Spatial Resolution

The phrase “subnational exposure and rapid damage assessment” requires:

clear naming of data sources,

geographic resolution (district/province/municipality),

and whether exposure data were updated dynamically or based on historical baselines.

The manuscript would benefit from clarifying how spatial mismatches were managed if outcomes were recorded at a different administrative level.

6. Figures and Visual Presentation

Some figure axes and labels (as seen in the PDF layout) appear truncated or insufficiently annotated.

Ensure each figure includes self-contained legends.

Add units and definitions for all plotted variables.

A flow diagram summarizing the data linkage pipeline (GDACS → exposure index → clinical data → wastewater operational events) would improve clarity.

7. Interpretation and Overgeneralization

The Discussion occasionally suggests broad applicability of the approach.

It would be appropriate to add one sentence acknowledging that:

GDACS alerts are global but wastewater infrastructure vulnerabilities are highly context-dependent.

8. Minor Editorial Issues

A few sentences would benefit from grammatical polishing and reduction of passive voice to improve readability.

Some terminology (e.g., “operational impacts,” “exposure levels”) should be used consistently throughout the text.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

To ensure your figures meet our technical requirements, please review our figure guidelines: https://journals.plos.org/plosone/s/figures 

You may also use PLOS’s free figure tool, NAAS, to help you prepare publication quality figures: https://journals.plos.org/plosone/s/figures#loc-tools-for-figure-preparation

NAAS will assess whether your figures meet our technical requirements by comparing each figure against our figure specifications.

Revision 1

Manuscript ID: PONE-D-25-49086

Title: Using GDACS to Anticipate Clinical and Operational Burden after Earthquakes: A Global Event-Level Analysis (2020–2024)

Journal: PLOS ONE

From: Ahmet Aykut, MD (corresponding author)

To: Dr. Anna Bernasconi, Academic Editor, and Reviewers

Dear Dr. Bernasconi and Reviewers,

We would like to thank you for the careful evaluation of our manuscript and for the constructive comments that helped us improve the clarity and transparency of the work. We have revised the manuscript accordingly and prepared a clean version and a tracked-changes version, as requested.

In brief, the main revisions are as follows:

1. We added a new main-text table (Table 1) summarizing all GDACS indicators used in the analysis, including units, operational definitions, thresholds, and rationale, and we updated Sections 2.4 and 2.8 to explicitly describe indicator selection and threshold choices.

2. We clarified the temporal structure of the analysis in Section 2.6, emphasising that all outcomes are event-level aggregates and that robustness is addressed through pre-specified sensitivity analyses on composite windows and outcome definitions.

3. We regenerated all figures according to PLOS ONE figure guidelines, with fully expanded axis labels, units, and self-contained legends, and we updated table and figure numbering consistently.

4. We added explicit legends for the Supporting Information tables (S1–S2) and aligned terminology (e.g. “operational burden/outcomes”, “GDACS-modelled population exposed within 100 km”) throughout the manuscript.

Below, we respond point-by-point to the reviewer’s comments. Reviewer comments are reproduced in italics.

1. Reviewer 2, comment 1 – GDACS Indicator Selection and Operationalization

“The manuscript mentions GDACS alerts and ‘subnational exposure’ but does not clearly describe which GDACS parameters were used (e.g., alert score, exposure index, population affected), how thresholds were defined, and whether the chosen thresholds have prior validation. Please add a concise table summarizing GDACS variables included, units, cut-offs, rationale for selection. This will increase transparency and facilitate replication.”

Response:

We thank the reviewer for this helpful suggestion to improve transparency and replicability. In the original submission, the GDACS indicators used (continuous GDACS score, alert level, moment magnitude, hypocenter depth, and GDACS-modelled exposed population) were described across Sections 2.3–2.5 but not summarised in a single location. To address this, we have added a dedicated main-text table that consolidates all GDACS indicators, their units, operational definitions, thresholds, and rationale: Table 1. GDACS indicators and operational definitions used in this study. This table specifies, for each indicator, the GDACS source/origin, how it is used in the analyses (descriptive, correlational, regression), any inclusion thresholds or windows applied (e.g. restriction to Orange/Red alerts, composite clustering windows, exposure-normalised mortality subset), and the justification and validation status of these choices.

In Section 2.4 (Variables and Linkage), we revised the opening description of GDACS variables to make the indicator set explicit and to refer the reader to Table 1. The relevant paragraph now reads: “All covariates and outcomes used in the analysis were derived from the GDACS alert feed and its hazard/exposure metadata, including the continuous GDACS score (0–10), alert level (Orange/Red), moment magnitude (Mw), hypocenter depth (km), and GDACS-modeled population exposed within 100 km, and were linked to the representative event (Table 1). Event-level deaths were abstracted from WHO/OCHA and national situation reports, consolidated using a fixed priority order (WHO/OCHA > national authorities > consolidated media), and linked to the same representative event.” This makes the complete set of GDACS-based indicators and their linkage to events fully explicit in the Methods.

With respect to thresholds and their prior validation, we have clarified these points in Section 2.8 (Methodological Considerations – Bias and Assumptions). Under “Selection and case definition”, we now explicitly state that the study universe was restricted to Orange and Red alerts in line with GDACS’ own categorisation of potentially damaging events, and that this restriction was specified a priori rather than being data-driven. We further clarify that the temporal (±24 h) and spatial (≤150 km) windows used to cluster alerts into composite events were pre-specified pragmatic choices informed by common seismic clustering practice, rather than thresholds previously validated for predicting clinical outcomes. Finally, we note that the robustness of these choices was evaluated through pre-specified sensitivity analyses using alternative temporal (±12–48 h) and spatial (100–200 km) windows, with details reported in Supplementary Table S1. We believe these additions, together with the new Table 1, address the reviewer’s concerns and substantially improve the transparency and reproducibility of the GDACS indicator operationalisation in our study.

2. Reviewer 2, comment 2 – Temporal Alignment Between GDACS Signals and Outcomes

“The methods should clarify the exact temporal resolution used when linking GDACS alerts to: clinical service indicators, wastewater operational disruptions. It is unclear whether you used: same-day correspondence, lagged effects (e.g., 1–3 days), cumulative exposure windows. Clarifying the temporal structure and providing justification (e.g., sensitivity analyses) will strengthen the analytical foundation.”

Response:

We appreciate the reviewer’s focus on temporal structure. The present study is conducted entirely at the composite earthquake–event level rather than at a daily time-series resolution. Our outcomes are limited to (i) field-hospital deployment (yes/no) and (ii) mortality (total deaths, and deaths per 100,000 exposed in a subset). We do not include wastewater or other infrastructure-related operational outcomes in this analysis; we apologise for any confusion caused by the framing of “operational burden” and now emphasise more clearly that our outcome set is restricted to health-service deployments and deaths.

Regarding temporal alignment, all GDACS indicators and outcomes are linked at the level of the composite event, which is defined by the pre-specified ±24 h / ≤150 km clustering window described in Section 2.3. Field-hospital deployment and mortality are treated as cumulative quantities associated with each composite event rather than as daily counts. We do not estimate same-day, 1–3 day lagged, or sliding cumulative windows across days because outcome data are only available as event-level aggregates from WHO/OCHA and national situation reports. To make this explicit, we have revised the Analytical Note (Section 2.6) to state: “All analyses were performed at the representative-event level; outcomes were defined as cumulative field-hospital deployment (yes/no), total reported deaths, and, in a subset, deaths per 100,000 exposed for each composite event. We did not model daily trajectories or lagged effects, because all outcome data were available only as event-level aggregates. Sensitivity analyses using alternative composite windows and outcome definitions are reported in the Supplement.” This clarifies that the temporal resolution of the analysis is the earthquake event itself, and that robustness is addressed through pre-specified sensitivity analyses on the composite windows rather than through day-level lag structures.

3. Reviewer 2, comment 3 – Definition of Wastewater “Operational Impacts”

“The outcomes described as ‘operational wastewater impacts’ appear broad. Please specify: whether these represent flow interruptions, contamination events, pump failure, overflow, staffing disruptions, or other categories, how each outcome was coded, and whether events of different severity were analyzed separately or pooled. A short operational glossary would be very helpful.”

Response:

We thank the reviewer for this detailed comment. The present study, however, does not include wastewater or other infrastructure-related outcomes. Our operational outcomes are restricted to health-service indicators—specifically, field-hospital deployment (yes/no) and mortality (total deaths, and deaths per 100,000 exposed in a subset of events). We do not collect, code, or analyse data on wastewater flow interruptions, contamination events, pump failures, overflows, or staffing disruptions, and therefore a glossary of wastewater impact categories is not applicable to this analysis. For this reason, no changes have been made to the manuscript text in response to this comment.

4. Reviewer 2, comment 4 – Statistical Modelling and Uncertainty Reporting

“The statistical section briefly mentions associations but does not detail: model type (e.g., logistic regression, GLM, Poisson, time-series), covariates included, handling of missing data, validation approach. Consider explicitly reporting: effect estimates with 95% confidence intervals, model diagnostics, and sensitivity checks (e.g., excluding borderline GDACS events). These additions would substantially improve interpretability.”

Response:

We appreciate the reviewer’s focus on statistical transparency and uncertainty reporting. In the current version of the manuscript, we already provide detailed information on model type, covariates, missing data handling, and internal validation, as well as effect estimates, confidence intervals, model diagnostics, and pre-specified sensitivity analyses. We summarise these elements and their locations here for clarity.

First, the model types used are explicitly described in Section 2.2 (Statistical analysis). All analyses are conducted at the composite event level. Associations between GDACS indicators and outcomes are quantified using Spearman’s rank correlation. For death counts, we use negative binomial regression to account for overdispersion, and results are reported as incidence-rate ratios (IRRs) with 95% confidence intervals. For field-hospital deployment (yes/no), we use a Firth penalised logistic regression, motivated by the limited number of deployment events and near-separation; this is introduced in the primary objectives and in the description of the analytical environment, and summarised in the model diagnostics table (Supplementary Table S2, Panel A).

Second, the covariates available and how they enter the models are detailed in Section 2.4 (Variables and Linkage) and summarised in the new Table 1 (“GDACS indicators and operational definitions used in this study”). The main predictors are the continuous GDACS score, alert level (Orange/Red), moment magnitude, hypocentre depth, and, in a subset, GDACS-modelled population exposed within 100 km. For rate-based analyses of deaths, exposure enters as a denominator (or offset), as specified in the Statistical Analysis section and co-primary endpoint description. We do not fit highly parameterised multivariable models beyond this parsimonious set, given the modest sample size and event counts.

Third, handling of missing data is described under “Missing data” in Section 2.8 (Methodological Considerations – Bias and Assumptions). We use complete-case analysis per model, and variable-wise missingness counts are tabulated in Supplementary Table S2 (Panel B). As noted in the Discussion, the amount of missingness is modest and was judged insufficient to warrant multiple imputation given the limited sample size and the primarily exploratory nature of the deployment model.

Fourth, internal validation and model diagnostics are reported in both the Methods and Supplementary material. For the field-hospital deployment model, discrimination (area under the receiver-operating characteristic curve, AUC), overall accuracy (Brier score), and calibration slope/intercept are specified in the primary objective and fully reported in Supplementary Table S2 (Panel A). In the Results, we highlight that the model achieves excellent discrimination (AUC ≈ 0.98) with acceptable calibration, acknowledging that this reflects internal performance only and does not constitute external validation. For the death models, we report IRRs with 95% confidence intervals and evaluate model fit via overdispersion checks and residual inspection.

Fifth, effect estimates with 95% confidence intervals are already presented for the regression models. For death counts, IRRs and their 95% confidence intervals are reported in the regression table and referenced in the Results. For the deployment model, we focus on predictive performance metrics (AUC, Brier score, calibration) rather than on the magnitude of the odds ratio, because the primary inferential interest is in predictive utility rather than in a fully adjusted causal effect estimate. Nonetheless, all model coefficients with standard errors and confidence intervals are available in the shared code and output.

Finally, sensitivity analyses are pre-specified and reported in Supplementary Table S1. These include (i) analyses using raw (non-composite) GDACS events, (ii) an alternative definition of deployment (broader temporary medical facilities), and (iii) alternative composite windows (±12/±48 hours; 100–200 km). Together, these robustness checks address concerns that the main associations might be driven by specific clustering choices or by the stricter deployment definition. We did not add an additional, ad hoc sensitivity analysis excluding “borderline GDACS events”, because this would further reduce effective sample size and overlap conceptually with the existing pre-specified robustness checks.

In view of the above, we did not modify the underlying models, but we have ensured that the Statistical Analysis, Variables and Linkage sections, and the new Table 1, as well as Supplementary Tables S1–S2, clearly document model types, covariates, missing data handling, internal validation metrics, effect estimates, and robustness analyses. We hope this summary clarifies the analytical structure and addresses the reviewer’s concern about interpretability.

5. Reviewer 2, comment 5 – Data Sources and Spatial Resolution

“The phrase ‘subnational exposure and rapid damage assessment’ requires: clear naming of data sources, geographic resolution (district/province/municipality), and whether exposure data were updated dynamically or based on historical baselines. The manuscript would benefit from clarifying how spatial mismatches were managed if outcomes were recorded at a different administrative level.”

Response:

We thank the reviewer for this comment. The current analysis does not use any subnational administrative exposure datasets or rapid damage assessment products. All predictors and outcomes are defined at the level of composite earthquake events. Specifically, as detailed in Sections 2.3–2.4 and summarised in Table 1, the GDACS indicators used in the models are the continuous GDACS score, alert level (Orange/Red), moment magnitude, hypocentre depth, and, in a subset of events, the GDACS-modelled population exposed within 100 km of the epicentre. Outcomes are limited to field-hospital deployment (yes/no), total reported deaths, and deaths per 100,000 exposed where an explicit exposure estimate is available. These quantities are linked to each composite event rather than to district-, province-, or municipality-level administrative units, and we therefore do not perform spatial joins across mismatched administrative levels.

The phrase “subnational exposure and rapid damage assessment” appears in the Abstract, Discussion, and Conclusion solely to indicate potential future directions—namely, that integrating GDACS hazard signals with additional subnational exposure maps, damage assessments (e.g., remote-sensing products), and health-system status indicators could, in principle, improve the alignment between alerts and expected clinical burden in subsequent work. These data layers were not available in a globally consistent form for the 2020–2024 earthquake cohort analysed here and were not used in the present models. For this reason, there are no additional data sources, geographic resolutions, or spatial mismatch procedur

Attachments
Attachment
Submitted filename: Point_by_Point.docx
Decision Letter - Anna Bernasconi, Editor

Using GDACS to Anticipate Clinical and Operational Burden after Earthquakes: A Global Event-Level Analysis (2020–2024)

PONE-D-25-49086R1

Dear Dr. Aykut,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. For questions related to billing, please contact billing support.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Anna Bernasconi, PhD

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Dear authors, the revised manuscript now meets PLOS One quality standards. We appreciate the improvements made with respect to the original version. The paper can be accepted in its current form.

Reviewers' comments:

Formally Accepted
Acceptance Letter - Anna Bernasconi, Editor

PONE-D-25-49086R1

PLOS One

Dear Dr. Aykut,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS One. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

You will receive an invoice from PLOS for your publication fee after your manuscript has reached the completed accept phase. If you receive an email requesting payment before acceptance or for any other service, this may be a phishing scheme. Learn how to identify phishing emails and protect your accounts at https://explore.plos.org/phishing.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Anna Bernasconi

Academic Editor

PLOS One

Open letter on the publication of peer review reports

PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.

We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.

Learn more at ASAPbio .