Peer Review History

Original SubmissionJuly 24, 2021
Decision Letter - Roland Bouffanais, Editor

PONE-D-21-24064Using Machine Learning to Emulate Agent-Based SimulationsPLOS ONE

Dear Dr. Angione,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Nov 06 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.
  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.
  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Roland Bouffanais, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Thank you for stating the following in the Acknowledgments Section of your manuscript: 

"ES is part of the Complexity in Health Improvement Programme supported by the Medical Research Council (MC UU 00022/1) and the Chief Scientist Office (SPHSU16). CA would like to acknowledge the support from UKRI Research England’s THYME project, and from the Children’s Liver Disease Foundation. This work was supported by UK Prevention Research Partnership MR/S037594/1, which is funded by the British Heart Foundation, Cancer Research UK, Chief Scientist Office of the Scottish 

Government Health and Social Care Directorates, Engineering and Physical Sciences Research Council, Economic and Social Research Council, Health and Social Care Research and Development Division (Welsh Government), Medical Research Council, National Institute for Health Research, Natural Environment Research Council, Public Health Agency (Northern Ireland), The Health Foundation and Wellcome"

We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. 

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows: 

"project, and from the Children's Liver Disease Foundation.  This work was supported by UK Prevention Research Partnership MR/S037594/1, which is funded by the British Heart Foundation, Cancer Research UK, Chief Scientist Office of the Scottish Government Health and Social Care Directorates, Engineering and Physical Sciences Research Council, Economic and Social Research Council, Health and Social Care Research and Development Division (Welsh Government), Medical Research Council, National Institute for Health Research, Natural Environment Research Council, Public Health Agency (Northern Ireland), The Health Foundation and Wellcome.

The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript"

Please include your amended statements within your cover letter; we will change the online submission form on your behalf

3. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match. 

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section

Additional Editor Comments (if provided):

Both reviewers highlight the fact that your manuscript has merit and that it deals with an important topic.

However, as you'll read below, both reviewers expressed some concerns and felt that some points should be addressed before this manuscript can be published.

We look forward to receiving your revised manuscript.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: N/A

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors propose a careful and detailed comparison of machine-learning surrogates emulating the behavior of an agent-based model designed to evaluate social care policies in the UK. The authors consider both accuracy of prediction and efficiency (i.e. speed). The use of emulators is sometimes the only way to perform sensitivity and calibration of ABMs: therefore, comparing various approaches and disclosing the codes needed to obtain such emulators is relevant. In this respect, the paper is very welcome.

Indeed, I feel that, upon adequate revision, the paper has the potential to make a solid contribution to the social simulation/ABM literature.

Below I provide my comments, not in order of importance

1. The authors claim that machine-learning methods have rarely been employed in the ABM literature (pag. 3). This is partially true, but I suggest them looking at the recent survey by Dahlke et al 2020. In addition, please note that Lamperti et al. 2018, which the authors cite, conduct an exercise which is similar to the one proposed in this paper by comparing surrogate models obtained with GP and Boosted Trees (through XGBoost) across different metrics and using various sample sizes. I suggest the authors to discuss their results in light of those obtained in Lamperti et al. 2018 and in the rest of the literature. I feel this paper consolidates - and substantially expand - some evidence which was already there: this is a value added.

2. One crucial element in the analysis of stochastic ABMs embedding a time component is ergodicity, which I intend here as the presence of a single statistical equilibrium to which various runs (i.e. model runs with different seeds and same parameter configuration) approximately converge after a warm-up period (see for example Grazzini 2012; Vandin et al. 2021). My feeling is that surrogate modeling needs ergodicity as a prerequisite. It seems to me that the model considered by the author is deterministic (apart from the choice of initial conditions); however, several ABMs embed random draws during the simulation, and I feel the authors need to discuss whether their exercise and results can be extended to this class of models and under which conditions.

3. Is the interpretability of the ML surrogate valuable? The authors acknowledge that surrogates might be useful not just to improve the efficiency of simulations (which is necessary for - e.g. - global sensitivity analysis), but to help understand the model behavior, which is often complex (pag.4). However, when discussing their results, they do not mention the issue of interpretability, which I feel as relevant instead. For example, decision trees or random forests might help detect which parameters are more relevant to obtain behaviors aligned with the actual model or to models’ fit with the empirical data, therefore helping understand the model itself and its “sensitive” parts. To the eyes of a modeler this is fundamental for calibration and sensitivity analysis. To the contrary, ANNs can difficulty offer such information. I would suggest the authors to build on their observation of pag. 4 and discuss this issue in their Discussion and Conclusions sections.

4. Figure 3a is relevant. It shows that GP emulators (as implemented by the authors) have difficulties. I feel it would be relevant to show (i) the ground truth to which the emulator should be compared and (ii) similar charts for the other surrogates. Also, I would add a note of caution; if the ground truth behavior (i.e. how the actual ABM responds to parameter) is relatively smooth or even linear, learning should be easier. I wonder how the different methods perform when facing highly non-linear model dependence on parameters. I suggest providing – at least - some discussions.

5. This is probably the main concern I have. I have issues at understanding why the accuracy of surrogate models - e.g. GP and Boosted trees - vary so largely when increasing sample size, especially from 800 samples to 1600 samples. The performance seems rather consistent from 200 to 800 samples, and then it drops quite dramatically for 1600 samples. This is rather counter-intuitive to me. What happens, e.g., at 1200 samples? Can they author provide a convincing explanation of this issue, which greatly affects the results proposed by the paper.

Refs

Dahlke, J., Bogner, K., Mueller, M., Berger, T., Pyka, A., & Ebersberger, B. (2020). Is the juice worth the squeeze? machine learning (ml) in and for agent-based modelling (abm). arXiv preprint arXiv:2003.11985.

Lamperti, F., Roventini, A., & Sani, A. (2018). Agent-based model calibration using machine learning surrogates. Journal of Economic Dynamics and Control, 90, 366-389.

Grazzini, J. (2012). Analysis of the emergent properties: Stationarity and ergodicity. Journal of Artificial Societies and Social Simulation, 15(2), 7.

Vandin, A., Giachini, D., Lamperti, F., & Chiaromonte, F. (2021). Automated and Distributed Statistical Analysis of Economic Agent-Based Models. arXiv preprint arXiv:2102.05405.

Reviewer #2: This paper presents a set of experiments to capture agent-based simulation input-output relationships using machine learning approaches. It does so by simulating a model called Linked Lives with different parameter settings. Many machine learning techniques are compared with the Gaussian process (GP) emulation technique.

This paper is written in a clear language and as a reviewer, I had an easy time following its structure. Sharing code and data is a good practice. Well done! Overall, I would like to see this paper published but there are some additional tasks that need to be conducted in order to get to that level. Here are some major concerns:

1. Most agent-based models are stochastic, meaning each run will lead to different output values even if input values are the same. The approach in this paper does not capture uncertainty generated by multiple runs of the same configuration.

2. would recommend finding the best performing ML technique first. Then, I would compare it with GP and other surrogate modeling techniques. Speaking of other techniques, Kriging is also a popular surrogate modeling technique to compare against.

3. Experiment sample size is quite limited. There is only one ABM used and there is only one output value observed. More ABMs (check Comses) and multi outputs can generate interesting and more comprehensive results.

Here are some minor points:

1. I believe the term “emulate” is not the right one to use in this context. Emulate indicates being “in place of” something and replacing it. In the context mentioned here there is “approximation” not replacement of the real simulation model. Even though the Gaussian process paper uses the term emulator, it is not the right use here. Surrogate model sounds like a better option.

2. The authors define ABM as both “Agent-Based Modelling” and “Agent-Based Model.” These two are not interchangeable. Please select one and use it consistently.

3. The first paragraph in the second page of the article starts describing agent-based modeling without any references. Some introductory references are needed.

4. “When ABMs are highly complex, performing these kinds of analyses becomes both time- and cost-prohibitive, potentially leading some modellers to truncate these analyses or eliminate them entirely.” I would add the implications of this statement which is related to verification and validation.

“ABM-based modelling” redundant words.

5. The paper calls the case study as moderate-complexity ABM. It’s not clear how model complexity is measured. Are there any objective techniques to do so?

6. Why spider plot :) Barplot would work better.

7. In figure 5, is the dotted line y=x? If so, please mention.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Hamdi Kavak

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Revision 1

Please find attached our point-by-point response.

Attachments
Attachment
Submitted filename: PLoS Response to reviewers.docx
Decision Letter - Roland Bouffanais, Editor

Using Machine Learning as a Surrogate Model for Agent-Based Simulations

PONE-D-21-24064R1

Dear Dr. Angione,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Roland Bouffanais, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #2: Authors have done a great job in addressing the reviewer comments. I have no further concerns about the paper. Well done.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: No

Formally Accepted
Acceptance Letter - Roland Bouffanais, Editor

PONE-D-21-24064R1

Using Machine Learning as a Surrogate Model for Agent-Based Simulations

Dear Dr. Angione:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Professor Roland Bouffanais

Academic Editor

PLOS ONE

Open letter on the publication of peer review reports

PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.

We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.

Learn more at ASAPbio .