Peer Review History
Original SubmissionJuly 17, 2020 |
---|
PONE-D-20-22242 How predictability affects habituation to novelty? PLOS ONE Dear Dr. Yanagisawa, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Oct 24 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols We look forward to receiving your revised manuscript. Kind regards, David K Sewell Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. Please modify the title to ensure that it is meeting PLOS’ guidelines (https://journals.plos.org/plosone/s/submission-guidelines#loc-title). In particular, the title should be "specific, descriptive, concise, and comprehensible to readers outside the field" and in this case it is not informative and specific about your study's scope and methodology. 3. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information. Additional Editor Comments (if provided): Dear Dr Yanagisawa, Thank you for submitting your manuscript to PLOS ONE. I have now received reviews from two experts and have also read through the manuscript myself. You will see that the reviewers differ in their impressions of your work. Reviewer 1 identifies several major methodological and statistical issues, whereas Reviewer 2 points out several places where the writing and procedural details could be presented more clearly. In my own reading of the manuscript, I shared several of the concerns voiced by Reviewer 1, and think they are sufficiently serious to preclude publication of the work in its current form. I think additional data collection and potentially an expanded set of experimental conditions will be required to bolster confidence in the theoretical claims that you are making. I leave open the prospect of submitting a revision of the current work, noting that considerable effort will be required to successfully address the more serious issues identified by Reviewer 1. I will summarize here what I see to be the most critical issues. The reviewers provide more detailed comments that I urge you to consider and respond to in full, should you consider revising the manuscript for resubmission. The most pressing concern with the manuscript is the very small sample size—only 8 participants in total, with 7 producing viable data for EEG analysis. Reviewer 1 notes that the sample size is unusually small, even for an ERP study, which raises serious concerns about the study’s capacity to accurately measure the habituation effects of interest. I think the only way to really address this issue is to collect additional data, ideally replicating the original study but with much larger sample. Reviewer 1 also raises concerns about the analysis of the surprise rating data—there is a clear floor effect for the congruent stimuli—which is very likely to be contributing to the key interaction effect described on pg. 11. The reviewer notes that adopting a different method of analysis might be a more appropriate way of analyzing the data, and I encourage you to pursue this. I also wonder whether it might be helpful to examine an “intermediate” level of congruency in an attempt to avoid floor effects on surprise. For example, could you present a modified sound produced by an instrument such that the instrument could not naturally that sound. (Perhaps applying something like a gated reverberation effect could create such an intermediate level of congruency, though feel free to tell me if this is off base.) A further concern with the data analysis lies with the binning. I found it a little unusual that the bins were so coarse-grained, given your interest in what appears to be a fairly rapid-onset habituation effect. A better strategy for tackling your key research question would seem to be to opt for more fine-grained aggregation to detect effects that arise early in the learning process. This issue is potentially entangled with the small sample size problem. Reviewer 1 suggests single-trial regression as a potentially more appropriate method of analysis for the kind of data you have. A final concern I had—also shared by Reviewer 1—is theoretical in nature. You provide a detailed quantitative overview of your Bayesian model, yet you investigate only qualitative predictions (e.g., an interaction effect on P300 amplitude). Finding these effects in the data does not selectively support the information gain hypothesis, and reliance on this kind of reverse inference is not justified. At minimum, there needs to be some consideration of alternative accounts for the data. Should you choose to revise and resubmit your work, I will send the manuscript back out for review, as there is a clear need for additional data collection. Yours sincerely, David Sewell [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: No Reviewer #2: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: No Reviewer #2: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Thank you for the opportunity to review this interesting manuscript by Ueda and colleagues. The manuscript reports an event-related potential study of habituation to novelty, using a format that combines both computational simulation and empirical work. The primary finding of the manuscript is of an interaction effect between participants' initial uncertainty regarding a sensory stimulus and the rapidity with which habituation occurs. This interaction is predicted by a Bayesian updating model, and the manuscript reports two pieces of evidence in support of this prediction: one involving self-reports of surprise, and one involving the amplitude of the P300 component of the event-related potential. In general, I thought that the general approach of the manuscript was sound, and the predictions from simulations were reasonable and sensible. However, I have several major concerns regarding the empirical approach of this manuscript. Of these, my feeling is that all but one might be addressed in a revision. Unfortunately, my feeling is that the remaining point, which concerns sample size and statistical power (point #1 below) severely undermines the interpretability of the empirical results reported in this manuscript. ## Critical point 1. The empirical section of this manuscript reports a sample of N=8 participants. After exclusion of 1 participant due to excessive eye-blink artefacts, therefore, the final results of this manuscript are derived from a sample of N=7 participants. The effect of this is that the manuscript is severely underpowered to estimate the effect size predicted by its simulations (between initial uncertainty and trial number). This problem is not unique to this manuscript; for instance, Button et al. (Nature Reviews Neuroscience, 2013) discuss the prevalence of underpowered studies in neuroscience more broadly. There are many consequences of underpowered studies, but the most important one for the present study is that significant results deriving from small samples are likely to significantly overestimate true effect sizes in the population. Notably, this study is underpowered even by the modest standards of ERP research; for instance, Clayson et al. (Psychophysiology, 2019) report the median sample size for ERP studies as 21 participants (3 times that of the sample in this manuscript). Unfortunately, short of replicating results in a larger sample, I believe that this represents a critical shortcoming of the manuscript that precludes interpretation of results. ## Major points 2. The results of the manuscript are presented as a test of a specific computational model of the P300 component of the ERP (in which information gain is operationalised as Kullback-Leibler divergence between posterior and prior beliefs). However, the test of this model presented in the manuscript is qualitative, rather than quantitative: in both the simulations and the empirical data, there is an interaction between trial number and initial uncertainty. However, such an interaction might also result from other cognitive mechanisms if we assume that their effect is to cause a roughly multiplicative decay in P300 amplitude over time. To give one example, an effect of participant fatigue might also produce the interaction between trial number and initial uncertainty reported in the manuscript. A far stronger test of the model's predictions would be, rather than discretising trials into 40-trial bins, to run a single-trial regression of P300 amplitudes onto information gain as measured by the manuscript's metric. If this model outperforms one in which P300 amplitude is predicted as a function of trial number (rather than KL divergence), this would constitute stronger evidence for the manuscript's conclusion. This would also be consistent with other recent information-theoretic and Bayesian approaches to analysis of the P300 component (see point 4 below). 3. An interaction between trial number and initial uncertainty is reported for participants' self-reported surprise ratings. However, I believe that this may be a statistical artefact resulting from the fact that self-report ratings of congruent items are at a floor level (see Figure 5). Indeed, from inspection of the self-report scores in the accompanying supplemental material, it appears that the overwhelming majority of participants reported surprise levels of '1' (no surprise) for these conditions. In this case it is not surprising that an interaction would occur with initial uncertainty, since surprise can hardly be expected to decrease any further for instruments high in initial certainty. I would also note that in this case, the assumption of the ANOVA (normality of residuals) is violated, and the manuscript might be better served employing a different analytic approach (e.g., probit regression; see Liddell & Kruschke, 2018, Journal of Experimental Social Psychology). 4. The manuscript's Bayesian model of the amplitude of the P300 is an interesting one, but it does not engage with a body of prior research on Bayesian models of the P300 component of the event-related potential. A Bayesian model of the P300 was first proposed by Kopp (2008; In M. K. Sun (Ed.), Cognitive sciences at the leading edge (pp. 87–96)), and a number of empirical papers testing this and other Bayesian P300 models have been published in the last 5 years (e.g., Kolossa et al., 2013, Frontiers in Human Neuroscience; Kolossa, Kopp & Fingscheidt, 2015, NeuroImage; Bennett et al., 2016, eNeuro; Kopp et al., 2016, Cognitive, Affective, and Behavioural Neuroscience; Bennett et al., 2019, Psychophysiology). The manuscript would be strengthened by engaging with these prior papers that prefigure the ideas proposed here. ## Minor points - Page 4, lines 86-87: "Accordingly, the information gain represents a decrease in uncertainty by experiencing an event". This is not quite correct; KL divergence measures the similarity between two probability distributions. These distributions may be quite dissimilar (i.e., high KL divergence) but still have equal uncertainty (i.e., entropy). For instance, a Beta(10, 1) distribution and a Beta (1,10) distribution have equal entropy but very different KL divergence. See also the equation of information gain with KL divergence on page 3, line 57. - Although the manuscript was well written in general, there were a number of grammatical and spelling lapses throughout. One prominent one is the title of the manuscript ("How predictability affects habituation to novelty?"). If the title is phrased as a question, this should be "How does predictability affect habituation to novelty"; if the title is not phrased as a question then "How predictability affects habituation to novelty" would be grammatical. - The derivations of the Bayesian update rule (Equation 6) and the information gain (Equation 8) are non-trivial, but only the results of these derivations (and not the process by which the equations are derived) are presented. Speaking personally, this made it rather difficult for me to check the logic of the manuscript's derivation. I would recommend including additional supplementary material detailing these derivations to aid the interested reader. Reviewer #2: 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript is technically sound. The authors describe a prediction derived from a previously published Bayesian model of novelty and report a single experiment to test that prediction. The results were consistent with the predictions of the model. 2. Has the statistical analysis been performed appropriately and rigorously? Appropriate statistical analyses are reported, but for ANOVA I would expect to see some indication of effect size estimates and associated confidence intervals. 3. Have the authors made all data underlying the findings in their manuscript fully available? Summary data are available. I would have liked the authors to give access to the stimulus videos also. It is possible to work out what they showed with reference to the methods and table, but the actual stimuli would help. 4. Is the manuscript presented in an intelligible fashion and written in standard English? The manuscript is written in standard English, and is generally intelligible, but I think that the very technical nature of the writing will make it difficult to access by all but a specialist audience. With relatively little effort. I am sure that the authors could make the manuscript accessible to a wider readership by simply explaining key terms and concepts in simpler ways and providing more detail. The concept of long-term novelty (page 1, line 38) might seem counterintuitive or even oxymoronic to some readers. The authors refer to predictability and uncertainty throughout the introduction of the manuscript, but these terms are first defined at the end of page six, and then in mostly mathematical terms. A clear behavioural/operational definition of predictability and uncertainty when the terms are first used would greatly improve the intelligibility of the manuscript. Clarity of expression could also be improved in places. The opening paragraph of the introduction, for example, states (lines 36-37) “Therefore, if one is experiencing unpleasant novel events, one should get used to them earlier; if one is experiencing pleasant novel events, one should be as unaccustomed as possible.” I assume that the authors mean that it would be desirable/advantageous for the individual to get used to novel unpleasant events as soon as possible to reduce their impact, but the way this sentence is written, it could be interpreted as meaning that novel unpleasant events lead to rapid habituation. In the procedure (3.1.3), the authors refer to eight videos, each shown 60 times, but then in the analysis section (3.1.5) mention four stimulus types and 120 exposures. It took me longer than it should have to work out that there were two videos of each stimulus type – a statement to this effect in the procedure would have helped. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
Revision 1 |
PONE-D-20-22242R1 How predictability affects habituation to novelty PLOS ONE Dear Dr. Yanagisawa, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by May 22 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, David K Sewell Academic Editor PLOS ONE Journal Requirements: Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice. Additional Editor Comments (if provided): Dear Dr Yanagisawa, Thank you for revising your manuscript. You’ll see that the Reviewers are both satisfied with the changes you’ve made to the manuscript, as am I. I am therefore happy to conditionally accept the manuscript, subject to addressing one outstanding issue, detailed below. Reviewer 1 flags one final point regarding their suggestion for conducting a single-trial regression analysis on the P300 data. I share their impression that noisy individual-level data does not necessarily preclude conducting the regression analysis. Given that you have performed the analysis, I think it would be prudent to report those results, even if they do not lead to definitive conclusions (e.g., if the disaggregated data are too noisy). It would be fine to report this analysis as a footnote to the primary analysis of the binned data. There, you could note any caveats or limitations of the single-trial analysis, also acknowledging its more exploratory nature. Of course, if I am overlooking any obvious reason why the single-trial analysis cannot be run, feel free to clarify this point. Yours sincerely, David Sewell [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: (No Response) Reviewer #2: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Thank you for the opportunity to review this revision of the manuscript by Ueda and colleagues. I appreciate the efforts that the authors have gone to to address my previous comments, and feel that my concerns have in large part been addressed by the substantial revisions in this new version of the manuscript. I have one lingering concern regarding the authors' response to my original point #2, in which I noted that splitting the data into 40-trial bins and analysing across bins represents a rather coarse-grained approach to testing model predictions, and suggested that the authors should employ a single-trial regression approach to allow a more fine-grained test of the model. In their response, the authors report that they did indeed conduct such an analysis, but that they do not report the results of the analysis because "the accuracy of estimation of P300 was worse than that of the conventional averaging method of P300 used in our study, so we could not test our hypothesis". I am slightly troubled by this response, which seems to side-step the substance of my recommendation. It is natural that single-trial measures of the P300 should have greater variance than the average of 40-trial bins, but I am afraid I do not quite understand why this variance should preclude running a regression analysis. My inclination is to request further details on this analysis from the authors, including a more detailed account of what features of the data prevented a single-trial regression analysis from being interpretable. However, I should state that I do not feel that this single issue in and of itself would make the difference between whether the manuscript should be published or not. I feel that the results are sufficiently compelling and rigorous in their current form, and would simply be interested to know more about the results of this regression analysis. I do not feel that this issue in and of itself is so major that it would preclude publication. In that light, I defer to the editor's judgment regarding this point. Reviewer #2: (No Response) ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
Revision 2 |
How predictability affects habituation to novelty PONE-D-20-22242R2 Dear Dr. Yanagisawa, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, David K Sewell Academic Editor PLOS ONE Additional Editor Comments (optional): Dear Dr Yanagisawa, Many thanks for your detailed responses to the comments and requests of the reviewers. I think the changes that you've made to the manuscript have greatly improved it, and am delighted to accept your manuscript for publication in PLOS ONE. Yours sincerely, David K Sewell Reviewers' comments: |
Formally Accepted |
PONE-D-20-22242R2 How predictability affects habituation to novelty Dear Dr. Yanagisawa: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. David Keisuke Sewell Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .