Peer Review History
| Original SubmissionMarch 26, 2024 |
|---|
|
PONE-D-24-07915Enhancing Machine Learning Performance in Cardiac Surgery ICU: Hyperparameter Optimization with Metaheuristic AlgorithmPLOS ONE Dear Dr. Ghousi, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Jun 30 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Xin Gao Academic Editor PLOS ONE Journal requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse. 3. We note that you have indicated that there are restrictions to data sharing for this study. For studies involving human research participant data or other sensitive data, we encourage authors to share de-identified or anonymized data. However, when data cannot be publicly shared for ethical reasons, we allow authors to make their data sets available upon request. For information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Before we proceed with your manuscript, please address the following prompts: a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially identifying or sensitive patient information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., a Research Ethics Committee or Institutional Review Board, etc.). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent. b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. Please see http://www.bmj.com/content/340/bmj.c181.long for guidelines on how to de-identify and prepare clinical data for publication. For a list of recommended repositories, please see https://journals.plos.org/plosone/s/recommended-repositories. You also have the option of uploading the data as Supporting Information files, but we would recommend depositing data directly to a data repository if possible. Please update your Data Availability statement in the submission form accordingly. Additional Editor Comments: Dear Author(s), Thank you for submitting your manuscript to PLOS ONE. We appreciate the effort you have put into this research and the potential impact it can have on ICU ventilator management. After carefully reviewing the feedback from the reviewers, I have decided that a major revision is necessary before we can consider your manuscript for publication. Given the mixed reviews, I have decided that a major revision is necessary. Please address the following key points to improve your manuscript: Provide clear explanations for your choice of models (CatBoost, LDA, ANN) and the use of simulated annealing over other methods such as cross-validation or grid search. Additionally, ensure that the rationale behind your hyperparameter tuning ranges is well-articulated. Reconsider the approach of removing features with more than 15% missing data. Explore and discuss imputation methods and assess the importance of these features to ensure they are not prematurely discarded. Validate your model across multiple independent datasets to enhance its robustness and generalizability. Address potential overfitting issues, particularly with ANN, and consider the use of ensemble models for improved predictive performance. (Optional) Include comparisons of simulated annealing with other optimization algorithms and techniques such as AutoML. This will help demonstrate the advantages and limitations of your approach. Discuss the clinical significance of your findings in more detail and incorporate aspects of model explainability to improve trust and adoption in clinical settings. Highlight how your model integrates with existing scoring systems and its potential impact on patient care. Ensure that all figures and plots in the manuscript are of high quality and clearly legible. Please provide a detailed response to the reviewers' and editor's comments and submit a revised version of your manuscript along with a point-by-point response letter. The revised manuscript will undergo another round of review to ensure the concerns have been adequately addressed. We look forward to receiving your revised manuscript and hope that you can address the issues raised to improve the quality and impact of your study. Best regards, [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Partly Reviewer #4: Yes Reviewer #5: No Reviewer #6: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: N/A Reviewer #4: Yes Reviewer #5: Yes Reviewer #6: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes Reviewer #4: No Reviewer #5: No Reviewer #6: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes Reviewer #4: Yes Reviewer #5: Yes Reviewer #6: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: it is a well-designed and executed study that addresses an important challenge in the healthcare domain. The combination of ML models with SA for hyperparameter optimization is an effective approach that has not been widely explored in the context of cardiac surgery ICU ventilator prediction. The enhancement in the predictive performance, particularly the increase in sensitivity (recall), after applying the Simulated Annealing algorithm for hyperparameter tuning is a noteworthy finding that can have direct implications for patient care and resource allocation in the ICU setting. To further strengthen your manuscript, I would suggest considering discussing the clinical significance of your findings in more detail, highlighting the potential impact on patient care and healthcare resource management. Also acknowledge the limitations of your study, such as the single-center nature of the dataset, and discuss the need for external validation of the proposed models. Lastly, explore the possibility of integrating your hybrid model with existing scoring systems (e.g., APACHE, SAPS) to further enhance the predictive capabilities and provide a more comprehensive assessment of patient risk. Reviewer #2: In the paper titled "Enhancing Machine Learning Performance in Cardiac Surgery ICU: Hyperparameter Optimization with Metaheuristic Algorithm," the authors present a significant advancement in predictive healthcare analytics. The study addresses the critical need for ventilator allocation in cardiac surgery ICUs, leveraging a hybrid approach that combines machine learning (ML) and deep learning (DL) models with a metaheuristic algorithm. One of the key strengths of this study lies in its innovative methodology, which employs Simulated Annealing (SA) for hyperparameter tuning, significantly improving the prediction performance of ML and DL models. Notably, the study demonstrates a remarkable 14% increase in sensitivity for Artificial Neural Networks (ANN) after applying SA. Furthermore, the authors provide comprehensive insights into the challenges of ICU care, particularly in ventilator management, and highlight the potential of ML techniques to address these challenges effectively. By leveraging a rich dataset obtained from a hospital, the study offers practical implications for optimizing ventilator allocation in critical healthcare settings. The paper's contribution extends beyond methodology, as it systematically addresses the gap in predicting ventilator needs specifically for cardiac surgery ICU patients. This focus is particularly relevant during pandemics or periods of increased ICU admissions, where efficient resource allocation can be lifesaving. Overall, the study provides a well-researched and meticulously executed analysis that not only enhances our understanding of ventilator management in cardiac surgery ICUs but also underscores the potential of ML techniques, especially when integrated with metaheuristic algorithms, to improve patient care outcomes in critical healthcare scenarios. Reviewer #3: Thanks for your contribution. I have couple questions below: Technical: 1. The paper conducted PCA but the feature space is not large - 32 features and PCA can throw unsatisfying results. Normally L1, L2 regularization can alleviate overfitting issue, all of the LDA/Catboost/ANN have ways to do that. 2. Do have concern over removing missing feature > 15%. Some may be important and useful in prediction. Have we tried any imputation method for missing values. Probably looking at feature importance to see if some missing features are badly removed. 3. The paper experimented with CatBoost/LDA/ANN but didn't give articulate the reasons behind it. 4. The paper didn't articulate the reasons using SA to conduct hyper-parameter tunings although its valid theoretically. Other methods, e.g. Cross-validation can do the tuning easily. 5. Have we checked overfitting issues when evaluation the models, ANN can be easily overfitted. 6. For predictive purpose, it may be benefitted from an ensembled model. 7. The paper lacks comparison of SA to other optimization algorithm. Publication: The plots in the end seem blurred. Reviewer #4: This study introduces a novel approach to predict the need for ventilators in cardiac surgery ICU patients by integrating machine learning (ML) and deep learning (DL) models with a metaheuristic algorithm, specifically Simulated Annealing (SA), for hyperparameter optimization. The research is highly relevant and valuable in reality, given the critical importance of ventilator allocation in healthcare settings, especially during the pandemic. Strengths: 1. The paper effectively combines various ML and DL techniques with a metaheuristic optimization algorithm, which is innovative in the field of medical data. 2. The topic is of high relevance, potentially offering significant improvements in ICU resource management and patient care. Weakness: 1. The study is limited to data from a single hospital source and the data access is restricted. Validation of the model across multiple independent datasets would enhance the robustness and general applicability of the findings. 2. There is insufficient discussion regarding the limitations of the applied models and the potential biases within the study. A thorough exploration of these aspects would provide a more comprehensive understanding of the model's utility and constraints. 3. In clinical settings, the explainability of predictive models is crucial for user trust and adoption. The study could be improved by incorporating aspects of model explainability and interpretability, particularly in explaining decisions in critical care scenarios. Reviewer #5: The problem stated in this paper is relatively simple in domain of machine learning. I did not deny its importance in the field of medical domain, but such a classification/regression problem (given data of a patient, determine if he/she needs ventilator) has been well studied, and many advanced techniques can be used to address this problem. Applying traditional machine learning algorithm such LDA, Catboost and ANN is nothing new. More advanced research has been using tabular transformers to address medical data like this. Regarding the Simulated Annealing (SA) to improve the ML and DL models' performance. I think a more realistic way of doing hyper parameter is doing grid search. The claimed performance improvement is questionable given the size of the dataset used in this paper (1092 records in total). The difference between each method, and before and after using SA is minor. Plus, i think in medical domain, sensitivity is more important than other metrics. assigning ventilator to a patient who does not need it might be a waster of resource, but not assigning a ventilator to a patient who actually need it will be life threatening consequence. If we take a look at the sensitivity of all model results provided in this paper, it's around 50%, which is more like a random guessing given this is a binary classification problem (need or not need a ventilator). Reviewer #6: Technically, I think this paper is well written. The algorithms and methodology are introduced clearly. I do have several problems from my perspective: 1. I think we do need some intuition of using certain ML mythology. Table1 introduces some of the ML method that other paper were using. It seems you are using some method that others not used before. But that is not sufficient to demonstrate the reason us using CatBoost, LDA or ANN. 2. If the focus for this paper is hyperparameter tuning, I think you need to explain why you choose such ranges for parameter tuning. 3. If we also want to compare between different models, we might want to make sure their search space are similar, or at least within the similar number of parameter combinations. It seems ANN's search space is bigger than other two. To this end, I think the comparison might not be fair. 4. There are many hyperparameter tuning methods. For example, AutoML works well if you want to compare between models. Can you compare the pros and cons of your method versus AutoML frameworks? In other words, why not using AutoML. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: Yes: Yuhe Zhang Reviewer #3: No Reviewer #4: No Reviewer #5: No Reviewer #6: No ********** [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 1 |
|
Enhancing Machine Learning Performance in Cardiac Surgery ICU: Hyperparameter Optimization with Metaheuristic Algorithm PONE-D-24-07915R1 Dear Dr. Ghousi, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Xin Gao Academic Editor PLOS ONE Additional Editor Comments (optional): Dear Authors, I am pleased to inform you that your manuscript, "Enhancing Machine Learning Performance in Cardiac Surgery ICU: Hyperparameter Optimization with Metaheuristic Algorithm," has been accepted for publication in PLOS ONE. The reviewers and I appreciate the efforts you have made in addressing the concerns raised during the review process. Your revisions have significantly improved the quality and rigor of your work, and we believe your research makes a valuable contribution to the field of machine learning applications in cardiac surgery ICU ventilator management. Specifically, you have: 1. Provided a more comprehensive discussion on the limitations of using a single-center dataset and the need for external validation using multi-center datasets to enhance generalizability. 2. Clarified your feature selection process, including justifications for the chosen scaling and upsampling methods. 3. Explained your cross-validation strategy and its importance given the sample size. 4. Corrected the reported metrics for the ensemble model, ensuring accuracy. 5. Strengthened the rationale for selecting specific machine learning models and their suitability for the classification problem at hand. 6. Addressed other specific concerns raised by the reviewers, improving the overall clarity and quality of the manuscript. However, there are some concern that should be addressed: please double-check the reported metrics for the ensemble model, as it is unlikely for all metrics (accuracy, precision, sensitivity, specificity, and F1-score) to have the same value. Ensure the accuracy of the reported results. While you have discussed the need for external validation, your study is still limited to a single-center dataset. In your future research, we strongly encourage you to prioritize multi-center validation to confirm the generalizability of your findings. We believe your work represents a significant contribution to the field and has the potential to inform clinical practice and inspire further exploration in this area. On behalf of the editorial team, I would like to congratulate you on your achievement and thank you for your dedication to this research. We look forward to the publication of your manuscript in PLOS ONE. Best regards, Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #4: All comments have been addressed Reviewer #7: (No Response) ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #4: Yes Reviewer #7: Partly ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #4: Yes Reviewer #7: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #4: No Reviewer #7: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #4: Yes Reviewer #7: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #4: Thank you for addressing my comments with detailed responses and revisions. Dataset Limitation: Your internal validation methods, including the use of ensemble learning and 10-fold cross-validation, are commendable and should strengthen the reliability of your findings. However, I still encourage future research to explore multi-center datasets when feasible to enhance generalizability. Model Limitations: Your explanation of the ensemble method and its impact on reducing overfitting and bias is well-articulated. Explainability: The integration of model explainability techniques, such as feature importance analysis, is a positive development. This addition should indeed help in bridging the gap between complex analytics and practical clinical application. Reviewer #7: The paper has proposed a new approach, which is ML-based to optimize ventilator allocation in cardiac surgery ICUs. The use of metaheuristic algorithms for hyperparameter tuning adds value to the existing body of research. Regarding the ML part of the paper, it explored multiple ML and deep learning models, including LDA, CatBoost, ANN, and XGboost, and used the weight-voting mechanism for the models. The paper is well-written Recognizing the paper's contribution to the prediction for surgery ICUs I do have some questions to this paper. - On page 7 regarding the feature selection, first, regarding the ID for each patient, if there’s no specific meaning of the ID, instead just like a random number, it might not be necessary to use that as one feature. Meanwhile, regarding the standardization of the numeric data, before doing the z-score scaling, it would be better to explore the data distribution first to show that z-score scaling is the most appropriate scaling method for the chosen datasets. This also applies to why use SMOTE to do the upsampling instead of using other up-sampling methods. - Some parts of the paper are lengthy, for example like the parts that explains the reason to split the dataset into train, validation, and test and the importance of validation datasets. Meanwhile, I wonder is there any cross-validation strategy used in splitting the datasets. Because the data sample size is kind of small, it might be better to apply some similar strategy to do data augmentation. - For table 4 ensemble model results, all metrics (Accuracy, precision, sensitivity, specificity and F1-Score) have given the same results 0.8169, wonder if there’s any error in this result because it is not likely that the model will have the same number across different metrics. - Regarding the ML technology is using (CatBoost, LDA, ANN and XGBoost), as the ML problem basically falls into classification, I wonder if there’s a particular reason for picking ANN and these boosting models. There could be choices of other models, and I feel the paper didn’t give strong reason or motivation for picking up these method. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #4: No Reviewer #7: No ********** |
| Formally Accepted |
|
PONE-D-24-07915R1 PLOS ONE Dear Dr. Ghousi, I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team. At this stage, our production department will prepare your paper for publication. This includes ensuring the following: * All references, tables, and figures are properly cited * All relevant supporting information is included in the manuscript submission, * There are no issues that prevent the paper from being properly typeset If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps. Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. If we can help with anything else, please email us at customercare@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Xin Gao Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .