Peer Review History
| Original SubmissionMarch 12, 2024 |
|---|
|
-->PONE-D-24-05900-->-->Modeling protective meningococcal antibody responses and factors influencing antibody persistence following vaccination with MenAfriVac® using machine learning-->-->PLOS ONE Dear Dr. Nasir, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Sep 19 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:-->
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols . We look forward to receiving your revised manuscript. Kind regards, Oyelola A. Adegboye, PhD Academic Editor PLOS ONE Journal requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse. 3. We suggest you thoroughly copyedit your manuscript for language usage, spelling, and grammar. If you do not know anyone who can help you do this, you may wish to consider employing a professional scientific editing service. The American Journal Experts (AJE) (https://www.aje.com/) is one such service that has extensive experience helping authors meet PLOS guidelines and can provide language editing, translation, manuscript formatting, and figure formatting to ensure your manuscript meets our submission guidelines. Please note that having the manuscript copyedited by AJE or any other editing services does not guarantee selection for peer review or acceptance for publication. Upon resubmission, please provide the following: The name of the colleague or the details of the professional service that edited your manuscript A copy of your manuscript showing your changes by either highlighting them or using track changes (uploaded as a *supporting information* file) A clean copy of the edited manuscript (uploaded as the new *manuscript* file) 4. Please include a complete copy of PLOS’ questionnaire on inclusivity in global research in your revised manuscript. Our policy for research in this area aims to improve transparency in the reporting of research performed outside of researchers’ own country or community. The policy applies to researchers who have travelled to a different country to conduct research, research with Indigenous populations or their lands, and research on cultural artefacts. The questionnaire can also be requested at the journal’s discretion for any other submissions, even if these conditions are not met. Please find more information on the policy and a link to download a blank copy of the questionnaire here: https://journals.plos.org/plosone/s/best-practices-in-research-reporting. Please upload a completed version of your questionnaire as Supporting Information when you resubmit your manuscript. 5. In the online submission form, you indicated that your data is available only on request from a third party. Please note that your Data Availability Statement is currently missing the name of the third party contact or institution / contact details for the third party, such as an email address or a link to where data requests can be made. Please update your statement with the missing information. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions -->Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. --> Reviewer #1: Partly Reviewer #2: Yes ********** -->2. Has the statistical analysis been performed appropriately and rigorously? --> Reviewer #1: I Don't Know Reviewer #2: I Don't Know ********** -->3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.--> Reviewer #1: No Reviewer #2: No ********** -->4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.--> Reviewer #1: Yes Reviewer #2: Yes ********** -->5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)--> Reviewer #1: The manuscript “Modeling protective meningococcal antibody responses and factors influencing antibody persistence following vaccination with MenAfriVac® using machine learning” by Nasir et al. addresses the public health challenge posed by meningococcal meningitis in the meningitis belt region in Sub-Saharan Africa. The authors argue that despite the introduction of the MenAfriVac vaccine to eliminate cases of Neisseria meningitidis serogroup A (NmA) in this region, the duration of post-vaccination immunity and the necessity of booster doses remains unclear. To address this gap, the authors propose a long and short-term model that uses demographic, medical and antibody variables to predict NmA antibody responses, as well as vaccine’s half-life. They use Explainable Boosting Machines (EBM) to model the long and short-term models and later they benchmark these results with linear regression and decision trees, where the EBM model shows better performance. The study is well motivated by the necessity of assessing post-vaccination immunity. It also proposes two interesting approaches (long and short-term) for predicting antibody kinetics. However, the manuscript has in my opinion some shortcomings: 1. Major issues 1.1. Introduction - The authors state the following: “Under the current study, we developed new computational models using machine learning techniques in order to improve the utility of modeling to inform decisions on PsA-TT scheduling and dosing”. It is unclear to me whether they refer to the approach/formulation of long and short-term models or rather to the EBM model. If they refer to the former, this is not really a new development of a computational model but a rather new formulation on how to predict antibody kinetics based on long and short term responses. If the authors refer to the latter, it should be stated more clearly that they have contributed to the development of the EBM tool. 1.2. Dataset - The exact variables which are available are unclear. Some demographic information is mentioned in the text and the authors refer to the supplementary information for the reader to check information on vaccination data. However, no such information was found in the supplementary S1. Referring to other variables as “etc” in the main text is not sufficient if an explicit list is not available in the supplementary. Table S1 contains information on variables incorporated into the long and short-term models, but as demographic information is missing from that table, it is unclear which data is really available for the authors. - There is no description of the cohort on the observables from the data collected (i.e., distribution of demographic variables, vaccination doses, clinical data). Descriptive statistics of the study variables will benefit the understanding of the cohort’s characteristics. 1.3. Modelling General - The formal equations of the model in the supplementary material S1 are only referred to at the end of the short-term model description in the main text. Having the reference earlier (i.e., when describing the long-term model) could enhance clarity of the model formulation. - The authors model each antibody task independently. This might lead to the loss of shared information between the tasks. - Antibody information often comes with a lower/upper threshold of the data. Is this the case here? If yes, the authors should consider censored regression using appropriate loss functions. Long-term modeling: - While figure 1 a-b illustrates well which antibody readings are used for the prediction task, the readability of the text could benefit from including the information that the model only uses the baseline antibody level and no further readings are included. - The authors claim that “A model limitation is that the estimate could be less accurate for small values of as the model is unlikely to encounter such time ranges with much frequency.” This should follow from the aforementioned sample description with descriptive statistics. Short-term modeling: - The authors state that “Prior studies have shown that the antibody level has an exponential relationship with time.[11] As a result, limiting the training samples to shorter ranges of time makes the model less prone to error, particularly if multiple readings from the same subject are available.” I agree that shorter time-intervals might reduce the variance of a prediction task but I do not see why this should be, relatively, higher for exponential relationships if the applied model allows for non-linear relationships. - The authors state that “However, the short-term model can be affected by compounding error if only the baseline antibody level 0 is provided.” I agree with this statement. I was wondering if this is the case for the author’s data (e.g. due to missingness) and if yes, how often is it the case? - For me the following formulation is unclear: “It is worth noting that the impact of the past vaccination history (before −1) is implicitly used by the model via the last antibody level −1, avoiding the need to explicitly include them, owing to the exponential modeling”. I agree that the use of the past vaccination is redundant as it is included in the last vaccination. However, it reads as if this fact is because of the exponential model - which is not clear to me. 1.4. Regression - The authors state that missing values are omitted for both long-term and short-term models; however, they do not justify this decision. I think it would be important to assess whether the dataset used for model training (i.e., without missing values) is a representative subsample of the original data. If the missing values follow a certain pattern, the model’s results might be biased to the characteristics of the training data and not generalizable to the whole dataset. - It is mentioned that the logarithm of antibody titer level is used for modeling. I believe this should be incorporated into the supplementary, such that the model’s specification follows: - It is mentioned that the dataset was divided into training and test sets only once, with a 80-20 split. However, it is not mentioned whether authors used an internal cross-validation on the training set for parameter tuning of the Explainable Boosting Machines (EBMs) model. I believe it is important to state how the EBM tuning was performed to enhance the reader's understanding on how the EBMs parameters were chosen. - Furthermore, splitting the data only once results in having one realization for the evaluation metrics, which are random variables due to their dependence on the sample. The robustness of the results could be enhanced by including outer-cross-validation. As the authors currently split the data by 80% and 20%, an outer 5-fold cross validation might be advisable. - The authors employ linear regression and decision trees as benchmark models. It is unclear to me why they did not consider using more complex models such as XGBoost or LightGBM as comparison models, as these are widely recognized for superior prediction accuracy over linear regression and decision trees. I believe it would be more interesting to compare the EBM to such more robust approaches to have a solid benchmark. - Furthermore, as the authors motivate their model as being different to the exponential model, it would be interesting to see how the exponential model performs on their data. 1.5. Half-life estimation - In this section, the authors mention that MenAfriVac as the first vaccination is the only one used for the half-life modeling. However, this decision is not further justified, so it is unclear to me whether this has a clinical reason or rather a reason from the structure of the dataset itself. - It is stated that individuals that do not reach the immunity threshold of 128 are excluded for the half-life analysis. As this means that the half life is only valid for individuals that reach this threshold, I suggest to mention that the half life is actually a half-life conditioned on a positive response to the vaccine. - The authors first speak about 99% Confidence Intervals and then about 95% Confidence intervals. Please clarify. - What does it mean that they bootstrapped the samples “and predicted their future antibody levels, incrementally adjusting the model by one day”? 1.6. Results - The authors evaluate and compare EBM against the benchmark models using the R-squared and RMSE metrics. They mention the following: “The EBM model performed better than both comparator regression models, both in long-term and short-term approaches, with respect to both metrics”. However, the evaluation of the model’s performance using both R-square and RMSE is redundant, as the R-Square is just a monotone transformation of the MSE. Thus, I would suggest only reporting one metric if the authors do not have a good reason to report both (in that case, I still suggest to report that both metrics favor the same models by construction). The statement “with respect to both metrics” is, however, in both cases misleading. - The authors make the following observation: “Finally, we found that using more samples resulted in the best performance, suggesting that more data might further improve our models in the future, irrespective of the study protocol”. I did not understand what is the purpose of testing the model with different amounts of observations, as it is well understood that using more available information will increase the model’s prediction accuracy. 1.7. Supplementary - In the section “Formulation of the machine learning approach from mixed effects model”, there is a typo in “[...] while Kl depends on Ab0 , the index of K should be L and not I. - It was not clear to me immediately whether the supplementary information on the model’s mathematical formulation refers to the long-term, short-term modeling, or a generalization of both models. I understand that is rather the latest, where depending on whether it is a long or short-term model, the term Abn will be different. However, I think that stating this explicitly and putting examples on how the formulation could change for the long and short-term model could enhance the reader’s understanding in this section. - The authors state: “Following White et al’s model, we used the log values of the antibody level, considering its exponential dependence on time. Another alternative was to take exponent of time as a feature while keeping the antibody level as is, however, this did not work well in our preliminary experiments; hence we limited our experiments to the first approach only.”. I think it could be interesting to explain further why it was that the exponential approach did not perform well in their model. This could improve the reader’s understanding on when to use log transformation or the exponential of time for modeling antibody responses. - Throughout the feature engineering section, there are strange characters in the text. This persists with different pdf viewers. - In the second point on the “Feature engineering” section, the authors mention that they created features using individual vaccine doses information. Specifically, they mention: “Some examples of original features associated with each dose are the type of the vaccine (PsA-TT or Hib-TT or something else), the dose amount, the time when it was administered, etc.”. I don’t think the authors should summarize which features were available for later feature creation with the term “etc”, as it is important for the reader to understand which variables were available to later create the new variables used for the models. Thus, I would suggest listing the full original features available. - Later on, they mention: “For the long-term model, we derived several features like the total dose of vaccines (both overall and PsA-TT only), the time of the most recent vaccine, the mean administering time of vaccines, etc.”. Again, the term “etc” should not be used, but instead the authors could refer to Table S1 which lists the features used for both long and short-term models. 1.8. General style - Mathematical expressions and symbols within text should be formatted appropriately, as they look like it is not a formula but rather plain text. 2. Minor issues 2.1. Introduction - In the sentence: “While adults get infected, children and adolescents do so at much higher rates”, a citation is needed to complement the statement. - The term medical variables is too broad; the variables are described later in the paper but could be good to already give a brief overview in the introduction (also because it is mentioned that this is one of the main differences between the authors’ approach and White et al.). 2.2. Half-life estimation - The sentence “For these data points, we used the initial antibody level (0) as input if it was above the threshold. If not, we used the highest recorded antibody level for the individual. Individuals with no recorded antibody levels above the threshold were excluded from the analysis” is a bit confusing. Maybe it is good to explain that if an individual does not record antibody levels above the threshold at any time point, it is excluded from the analysis. Reviewer #2: Major comments: • Overall this is an interesting manuscript using innovative methods to estimate duration of protection of MenAfriVac, which is an important public health and policy question that has been previously answered but can benefit from being revisited with newer methods. More detailed comments: Abstract: • Is there an extra word in this sentence? “In the short-term model, we found moderately high performance (R-squared = 0.59) for out-of-training-data subjects and END even better performance (R squared = 0.83) in the long-term evaluation.” Methods: • In the half-life estimation methods, “If an individual had no recorded antibody levels above the threshold, the subject was excluded from the analysis.” Can you explain this a bit more? I would imagine that individuals that don’t respond to the vaccine would be important to include in calculating/estimating population half-life, as they don’t help increase population immunity. If vaccine non-responders are excluded, then the estimated half-lives only pertain to those individuals that respond to the vaccine, so might be then artificially high when thinking about population-wide immunity. Could non-responders be considered to contribute 0 time to the half-life calculations, as they spent no time above the immunity threshold? Discussion: • It looks like there is a missing reference in the first paragraph. • WHO policy recommendations for the new lower-cost pentavalent conjugate vaccine are available at: Meningococcal vaccines: WHO position paper on the use of multivalent meningococcal conjugate vaccines in countries of the African meningitis belt, January 2024 • It would be interesting to also compare these estimated MenAfriVac duration results to estimates of duration of protection for other meningococcal conjugate vaccines Figures • Figure 4 – can the x-axis be labeled? ********** -->6. PLOS authors have the option to publish the peer review history of their article (what does this mean? ). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy .--> Reviewer #1: No Reviewer #2: Yes: Heidi M. Soeters ********** [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org . Please note that Supporting Information files do not need this step. |
| Revision 1 |
|
-->PONE-D-24-05900R1-->-->Modeling protective meningococcal antibody responses and factors influencing antibody persistence following vaccination with MenAfriVac® using machine learning-->-->PLOS ONE Dear Dr. Nasir, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.-->--> -->-->Please submit your revised manuscript by Mar 30 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:-->
-->If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols . We look forward to receiving your revised manuscript. Kind regards, Oyelola A. Adegboye, PhD Academic Editor PLOS ONE Journal Requirements: Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions -->Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.--> Reviewer #1: All comments have been addressed Reviewer #3: (No Response) ********** -->2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. --> Reviewer #1: Yes Reviewer #3: (No Response) ********** -->3. Has the statistical analysis been performed appropriately and rigorously? --> Reviewer #1: Yes Reviewer #3: (No Response) ********** -->4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.--> Reviewer #1: No Reviewer #3: (No Response) ********** -->5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.--> Reviewer #1: Yes Reviewer #3: Yes ********** -->6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)--> Reviewer #1: I appreciate the authors' detailed responses to my inquiries. Their revisions have effectively addressed all of my concerns and significantly improved the manuscript. I have only two minor suggestions remaining; once these are addressed, I believe the manuscript will be ready for publication. Minor Points 1) Original Comment Antibody information often comes with a lower/upper threshold of the data. Is this the case here? If yes, the authors should consider censored regression using appropriate loss functions. Author’s Response The lower limit of quantitation for the assay was 4. Anything below that was given the value of LLOQ/2, which equals 2. There was no upper limit. While we did not use censored regression as the current study focused on antibody levels as an effect of vaccination which rarely involved the lower limits, but it would be interesting to explore this as future work, particularly if the dataset and goal of the modeling involve many values below the threshold. New Comment: I agree with the authors that if censoring is limited, then using methods for censoring might not be necessary. However, I suggest to discuss this briefly in the paper and to mention the fraction of censored observations. 2) Original Comment The authors state that “However, the short-term model can be affected by compounding error if only the baseline antibody level 0 is provided.” I agree with this statement. I was wondering if this is the case for the author’s data (e.g. due to missingness) and if yes, how often is it the case? Author’s Response The statement was intended to suggest a real-world test case scenario instead of reflecting the data. If applied for a new study with only the baseline reading available, the short-term model might have compounding errors to predict antibody levels at multiple points of time in the future while using the predicted level at each point as baseline for the next prediction. New Comment: I suggest to include this reasoning in order to avoid confusions. Reviewer #3: The manuscript “Modeling protective meningococcal antibody responses and factors influencing antibody persistence following vaccination with MenAfriVac® using machine learning” explores the modeling of protective meningococcal antibody responses and factors influencing antibody persistence following PsA-TT (MenAfriVac) vaccination. The study employs machine learning techniques, specifically Explainable Boosting Machines (EBM), to develop both short-term and long-term predictive models. These models incorporate demographic, clinical, and serological data to estimate antibody kinetics and vaccine half-life. The work is well motivated, potentially being relevant in the context of optimizing vaccination strategies, well-structured and generally clear; however, I have some concerns and suggestions that, if addressed, could enhance the clarity, reproducibility, and overall transparency of the study. Major concern: - Authors should provide a more detailed description of the pre-processing steps applied to both predictive and outcome variables? While some information is provided in Supplementary 1, a more explicit explanation in the main text could improve clarity for the reader. Specifically: • Was batch correction applied to the outcome variable, given that measurements were obtained from different studies under potentially varying experimental conditions? If not, please clarify the choice • Were predictive variables subjected to normalization, decorrelation or other pre-processing techniques? Since the performance of some models used for comparison may be sensitive to these pre-processing choices, further details would be beneficial. Minor concerns: • The manuscript compares the model proposed by White et al with Linear Regression, Regression Trees, LightGBM, and XGBoost. Could the authors elaborate on the rationale for selecting these specific models? For example, were other ensemble or deep learning methods considered? A brief justification of these choices would provide a better context. • In the "Half-life estimation" section, the authors define a protective antibody threshold of 128 rSBA. Could they provide a reference or a more detailed explanation supporting this threshold? A brief discussion on why this value is considered appropriate would strengthen the validity of this assumption. • The "Model Interpretation" section presents feature importance rankings in Figures 4a and 4b, limited to the top 15 features. Would it be possible to provide a table or supplementary plot including all feature importance values? This would enhance transparency and allow readers to assess the relative contributions of less influential variables. • In the "Discussion" section, the phrase "we describe the development and evaluation of a machine learning algorithm to characterize..." might be somewhat misleading, as the original model was developed by White et al. A more precise wording, emphasizing the application and evaluation rather than the development of the model itself, would be more appropriate. • Supplementary 1: The subscripts a, l, and s are not clearly explained. A brief clarification would enhance readability. • Tables S1 and S2: Do these tables share the same legend? Please write a separate legend for each Table. • In the "Regression" section, the manuscript states that the long-term model includes 39 variables, while the short-term model includes 14. However, these numbers do not align with those in Table S2. Could the authors clarify whether this discrepancy arises from interaction terms, dummy encoding of categorical variables or something else? ********** -->7. PLOS authors have the option to publish the peer review history of their article (what does this mean? ). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy .--> Reviewer #1: No Reviewer #3: No ********** [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org . Please note that Supporting Information files do not need this step.--> |
| Revision 2 |
|
Modeling protective meningococcal antibody responses and factors influencing antibody persistence following vaccination with MenAfriVac® using machine learning PONE-D-24-05900R2 Dear Dr. Weeks, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Oyelola A. Adegboye, PhD Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions -->Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.--> Reviewer #1: All comments have been addressed Reviewer #3: All comments have been addressed ********** -->2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. --> Reviewer #1: Yes Reviewer #3: Yes ********** -->3. Has the statistical analysis been performed appropriately and rigorously? --> Reviewer #1: Yes Reviewer #3: Yes ********** -->4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.--> Reviewer #1: Yes Reviewer #3: Yes ********** -->5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.--> Reviewer #1: Yes Reviewer #3: Yes ********** -->6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)--> Reviewer #1: The authors have addressed my remaining comments and the manuscript is in my opinion ready for publication. Reviewer #3: (No Response) ********** -->7. PLOS authors have the option to publish the peer review history of their article (what does this mean? ). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy .--> Reviewer #1: No Reviewer #3: No ********** |
| Formally Accepted |
|
PONE-D-24-05900R2 PLOS ONE Dear Dr. Weeks, I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team. At this stage, our production department will prepare your paper for publication. This includes ensuring the following: * All references, tables, and figures are properly cited * All relevant supporting information is included in the manuscript submission, * There are no issues that prevent the paper from being properly typeset You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps. Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. If we can help with anything else, please email us at customercare@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Assoc Prof Oyelola A. Adegboye Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .