Peer Review History
| Original SubmissionMarch 3, 2025 |
|---|
|
PONE-D-25-08088MSMCE: A Novel Representation Module for Classification of Raw Mass Spectrometry DataPLOS ONE Dear Dr. Xiong, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Jun 20 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols . We look forward to receiving your revised manuscript. Kind regards, Hirenkumar Kantilal Mewada Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1.Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. Please note that funding information should not appear in any section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. Please remove any funding-related text from the manuscript. 3. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match. When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section. 4. When completing the data availability statement of the submission form, you indicated that you will make your data available on acceptance. We strongly recommend all authors decide on a data sharing plan before acceptance, as the process can be lengthy and hold up publication timelines. Please note that, though access restrictions are acceptable now, your entire data will need to be made freely accessible if your manuscript is accepted for publication. This policy applies to all data except where public deposition would breach compliance with the protocol approved by your research ethics board. If you are unable to adhere to our open data policy, please kindly revise your statement to explain your reasoning and we will seek the editor's input on an exemption. Please be assured that, once you have provided your new statement, the assessment of your exemption will not hold up the peer review process. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Partly Reviewer #3: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: No Reviewer #2: No Reviewer #3: I Don't Know ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: This manuscript presents MSMCE, a novel deep learning-based feature representation module designed for the classification of raw mass spectrometry (MS) data. The proposed method addresses known limitations of single-channel feature representations by introducing multi-channel embeddings with a residual-like concatenation strategy. The model is well-motivated, technically sound, and its performance is evaluated across four diverse and publicly available datasets using multiple neural network architectures (e.g., ResNet, DenseNet, LSTM, Transformer). The availability of code and data enhances the reproducibility and transparency of the study. The authors demonstrate notable empirical improvements in classification accuracy and computational efficiency, supported by clear ablation studies. However, the manuscript would benefit from several key improvements to strengthen its statistical and practical contributions: Major concerns: 1. While performance metrics (accuracy, precision, recall, F1) are thoroughly reported, the manuscript lacks statistical testing to validate the significance of the observed improvements. Repeated runs with different random seeds and the use of standard tests (e.g., Wilcoxon signed-rank test, bootstrapping confidence intervals) are recommended. This is particularly important for small-sample settings (e.g., Canine Sarcoma dataset). 2. The MSMCE module's effectiveness is well demonstrated quantitatively, but the biological meaning of the learned features is not explored. Given the biomedical context of mass spectrometry data, the manuscript would benefit from a discussion (or illustrative example) of how MSMCE embeddings relate to biologically meaningful features or clinically relevant patterns. This is important for building trust and enabling adoption in translational research. 3. The manuscript refers to a “residual connection” via channel concatenation, which is misleading, as no additive operation is applied. Please revise the terminology to avoid confusion with standard residual connections. 4. Details about hyperparameter selection, learning rate tuning, and batch sizes should be more explicitly stated or summarized in a table. 5. The methods for computing FLOPs and model size should be clearly described. 6. The t-SNE plots and radar charts in the supplementary material are helpful but not discussed in the main text. Please integrate interpretations of these visualizations into the Results or Discussion to show how MSMCE improves feature separability and efficiency. Minor concerns: • The manuscript is well-written in general. Minor edits are suggested for typographical consistency (e.g., consistent table formatting, figure references). Also, be precise with terms such as "residual connection." • The references are not cited in sequential order (e.g., the citation sequence jumps from [3–8] to [10], or from [24–26] to [22,32,39]), which may confuse readers and should be corrected for consistency. • Although ReLU and Dropout are standard components in deep learning, they should be briefly explained for clarity, given that PLOS ONE targets a multidisciplinary audience that may not be familiar with these concepts. • Figures 1 to 3 are of very poor quality, with unreadable text. They should be replaced with higher-resolution versions to ensure legibility and clarity. Reviewer #2: Authors have proposed a novel perspective and strategy for using feature-representations from convolutional neural network layers in analyzing complex, and spatially and temporally intertwined MS data. In the proposed MSMCE, the feature-representations from two CNN layer (referred to as channel) has been used to enhance data representation. This multi –channel representation, accompanying an encoder layer and a feature integration strategy has enhanced performance of the model and decreased computational cost of model training. Authors have compared their proposed model to existing models and have done an ablation study to show contribution of embedded channels. While the novel innovation and proposed technique enhances the performance of the model and opens a new perspective, the paper lacks major details and statistics necessary for scientific reporting. The most important missing aspect is proper randomization and testing of the model. It is not clear if the reported enhancement is significant in comparison to the existing models. From the ablation study (table 7), I am convinced that the multi-channel embedding improves the performance, but I have major concerns about how the results have been reported. A lot of details about the models, their hyperparameters, training and testing , and regularization strategies are missing, making it hard to properly evaluate and compare contribution for multi-channel embedding. Here are a few notes: The last sentence of the abstract: “Experimental results ...MS data classification.” The link between reduced computational resource and generalizability of the model is not given and trivial. This sentence in the introduction: “The large ... cancer detection”. The challenges need to be explicitly included. Explicitly including the challenges help to develop motivation of the paper. This sentence: “These operations ... MS signal acquisition,” the issues need to be explicitly listed: such as peak shift, etc This sentence: “Nevertheless, the high dimensionality of MS data combined with limited sample sizes makes it challenging for traditional methods to effectively meet the demands of data analysis [11].”, It is not the high dimensionality of data or small sample size that limits the performance of the traditional methods, but the mismatches between bathes and the need for preprocessing of data so the methods work. Actually, deep learning models need more data to be trained on compared to conventional methods especially if the data has higher complexity and dimensions. Reference 11 is not correctly referenced as it talks about using matrix factorization and using Bayesian framework to overcome such problems and does not explicitly relay the logic used in this sentence. Regardless, the sentence is not following logic and message of this paragraph. This sentence: “These methods often fail to fully exploit the latent information within the data, thereby limiting their overall performance.”. No reference. This claim is not accurate and somewhat controversial. The concern is: these methods (if used correctly) are robust as cited in the same paragraph, but preprocessing data to feed to these models require careful fine tunings. Needs to be rewritten, maybe restructuring like: Achieving optimal performance with these methods typically involves multiple stages of preprocessing and fine-tuning of model parameters to fully leverage the rich, latent information embedded in complex MS data. This sentence:” In recent years, deep learning has emerged as a dominant technology”. Deep learning is not a technology, rather an approach or methodology. The term channel in this article has been used in title and as a cornerstone of the article. For the first-time reader, the term is confusing, because, in the context of MS data, channel does not have a single universal definition. The authors are using channel as a concept of feature representations from CNN layers. This becomes clear later in the methods section. The term channel needs to be clarified and defined explicitly in the introduction. There is also some mix-up of term channel and concept of dimension. For example, in this sentence: “In the field of image classification, multi-channel images contain complementary information across different channels, enabling a more comprehensive description of the target compared to single-channel representations”. Consider confusion in this sentence: “These multi-channel features not only enrich feature representations but also enhance the model’s adaptability to high-dimensional data.” I suggest definition of term “channel” in the introduction, and then being careful not to use it in interchange with dimension- representation- feature-etc This sentence does not read well: MS2DeepScore [25] employs a Siamese neural network to learn low-dimensional embeddings of MS vectors for predicting the structural similarity between chemical compounds, which is also applicable to MS Clustering. Maybe a comma is missing after vectors? In this sentence: Studies have shown that deep learning can directly capture complex patterns”, what is the implication of the word “directly”? What would the indirect way be? I suggest removing the word “directly” deep classification model” is not a valid term. In the section where authors outline the contribution of this study: “End-to-End Training Framework”, training the DL+classifier model in an end-to-end manner is not innovative and is widely used. See: Seddiki, Khawla, et al. "Early diagnosis: End-to-end CNN–LSTM models for mass spectrometry data classification." Analytical Chemistry 95.36 (2023): 13431-13437., “Dimensional Adaptation of Multi-Channel Embedding Representations:” This contribution is not clear. It is not clear what authors mean by “adaptability”. “Compared to the original single-channel MS vectors, the embedded multi-channel vectors exhibit better adaptability within CNN architectures.”, how has the adaptability been quantified? What are the statistics of the enhanced adaptability, where is the pvalue? A general feedback to the authors. Authors have used a fully connected layer as the first layer. It is not wrong and authors’ choice to do so, but in my opinion, it is counter intuitive to use a flat layer to reduce dimensionality. The local MS information is lost this way. When transforming using WX+b, all local dependence is lost (for example peak shape, etc). This is counter-intuitive with using the CNN in the rest of the structure of the model. In most of the cited papers, usually a CNN layers comes first ensuring capturing the local dependencies of peaks, etc. But again, the proposed model by the authors is valid. E is not a “feature matrix” in the classic sense (i.e., not raw features you designed or extracted). It is more accurate to call it: A latent embedding, or A learned representation Figure 1: input is BxD but in the figure it is illustrated as 1-D Addition of channel dimension is poorly shown in the figure 1 making it harder to understand. The green input vector (which says embeding_dim) directly goes into 1D convolution skipping adding a dimension to. The convolutions are 1D, and the Eprime is Bx1xd, what is “3” in dimension of K1 (1xC/2x3), which dimension the convolution is happening? The same question for K2. It seems the notion and the mapping between the size of K1 and K2 and the output is inconsistent. “By transforming feature vectors into multi-channel representations, this module enhances adaptability for downstream classification tasks.” this claim needs to be proven in a quantifiable way. It is not clear what does adaptability mean. The training procedure depicted in figure 2 is in contrast with what has been mentioned as “End-to-End Training Framework”, Figure 2 is showing the backpropagation path back to only classifier module leaving out the MCMCE. “enabling an adaptive representation of the input.” the term “adaptive” has not been used correctly. 2.6 Data Processing Workflow: This section starts with a few paragraphs as its own introduction. These paragraphs need to be moved to the introduction or discussion depending on the context and in this section only the method should be discussed. The data processing section leaves most of the details out, referencing [2],[10]. details of the method should be outlined. Section 2.7- What is the strategy to divide data to training, validation, and test sets? Has k-fold strategy been used? How many times the model has been trained? What is the strategy to avoid overfitting? Are all the results related to a single (but same for all models) trial division? Or the random seed might be different for each tested model? In the table 2,3 and 5 there are values for precision and F1 that are below 0.5. This does not make sense, because any model that has been trained on data should be better than a totally random model with 50% performance. Also, a value of exactly 0.5000 for recall seems not to be correct considering variability in the MS data. Table 2 and 3 does not report number of trained models or any variability related to the performance. The absolute value shows enhancement when the MSMCE has been included. However, no significance statistical metric has been reported. The proper way would have been multiple trainings using multiple random divisions of trials and then reporting an average, std and a pvalue for each comparison. “representing an increase of 1.608%”, this enhancement is not of value, unless repeated over several randomizations and then averaged. Tables 2, 3,5 and 6 report performance for models LSTM and transformer, without providing details of architecture of each of these models. LSTM layers or transformer architecture could be used in a DL-based pipeline in many ways. Details need to be properly reported. Table 6 (class12), precision and recall reported for multi-class arrangement. It is not clear if the final precisions the average of precisions for each class or a different strategy has been used. Figure 4, the reported accuracy drops at some epochs, which makes the reported accuracy metric to be questionless considering optimizers like Adam and loss functions like cross-entropy. How do the authors justify the drops? ResNet-50, DenseNet-121, and EfficientNet-B0 should be properly referenced. Figure 5, the lines should be transparent (alpha=0.8) for better visibility. The lines are blocking each other. Reviewer #3: I would like to begin by congratulating the authors for this very interesting and well-executed piece of work. This well-written article introduces a novel spectral representation technique that replaces traditional pre-processing methods with a neural network model employing multi-channel embedding. The model is clearly presented and supported by a well-structured introduction. The limitations of the approach are adequately discussed in both the Materials and Methods section and the Discussion. However, I have several suggestions to enhance the clarity and completeness of the manuscript: Image Quality: The figures are currently of insufficient quality, making them unreadable. This significantly hinders the reader's ability to follow the results. I strongly recommend improving the resolution and clarity of all figures to ensure they are legible and informative. Description of the Data: More detailed information about the dataset is necessary. Readers unfamiliar with this data may not understand its specific challenges, particularly in the context of classification tasks. Please elaborate on the nature of the dataset, potential difficulties, and how these may relate to the performance of the proposed representation technique. Error and Robustness Metrics: While the provided metrics illustrate the model’s performance, the lack of error analysis is a limitation. Including confidence intervals or other measures of uncertainty would provide a clearer picture of the method’s robustness. It would also facilitate a more rigorous comparison between methods. Comparison with Traditional Methods: The proposed technique is positioned as an improvement over traditional feature engineering methods. However, the manuscript does not clearly specify which traditional techniques were used to generate the “Original” baseline. Please provide more details about these baseline methods so readers can better assess the added value of MSMCE. Once again, I thank the authors for their contribution, and I also thank the editors for giving me the opportunity to review this manuscript. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean? ). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy . Reviewer #1: No Reviewer #2: Yes: Amir Akbarian Reviewer #3: No ********** [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org . Please note that Supporting Information files do not need this step. |
| Revision 1 |
|
PONE-D-25-08088R1MSMCE: A Novel Representation Module for Classification of Raw Mass Spectrometry DataPLOS ONE Dear Dr. Xiong, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Aug 03 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols . We look forward to receiving your revised manuscript. Kind regards, Hirenkumar Kantilal Mewada Academic Editor PLOS ONE [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed Reviewer #2: (No Response) Reviewer #3: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: No Reviewer #3: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The authors have provided a comprehensive and thoughtful revision. All reviewer concerns were addressed appropriately and with substantial improvements to both the clarity and rigor of the manuscript. Key contributions such as the statistical validation of performance improvements, clarification of convolutional architecture and terminology, and enhanced methodological detail, strengthen the manuscript considerably. Reviewer #2: The reported p-values appear inconsistent and potentially incorrect. All five instances report the exact same value (p = 0.03125), regardless of the magnitude of performance improvement. For example, in Table 5, the accuracy of the Transformer model on the Canine Sarcoma dataset increases markedly from 0.78 ± 0.00 to 0.99 ± 0.01 — a substantial improvement that would typically yield a much lower p-value (e.g., p < 0.01) if statistical significance were properly assessed. The uniformity of the reported p-values across different metrics and datasets raises concerns about the validity of the statistical testing procedure. Clarification is needed on how these p-values were computed and whether appropriate statistical tests were applied in each case. As noted in the previous round, the performance values reported in Tables 2, 3, and 4 — such as 0.25, 0.33, and 0.4 — are notably lower than what would be expected from a random classifier in a binary classification task. This suggests that the number of model training iterations may have been insufficient to yield a reliable evaluation. The authors report using 6-fold cross-validation (k=6); however, this was applied to a single 90/10 train-test split, meaning the same data partitioning was used throughout the evaluation (i.e., n=1). Given the class imbalance evident in the t-SNE plots provided in the Supplementary Materials, repeating the 90/10 split multiple times is essential to ensure a fair comparison between the baseline and enhanced models. Each repetition can be followed by k-fold cross-validation (potentially with a lower k if computational feasibility is a concern). This approach helps mitigate the impact of any train-test split, smoothing out performance fluctuations caused by randomness in the data. For instance, a low score such as 0.25 from one split may be balanced by higher scores in others, leading to a more representative average. The p-values in the tables appear to be reported selectively and not consistently across all model comparisons. Specifically, only three p-values are provided: the accuracy of ResNet-50 on the NSCLC dataset, the accuracy of DenseNet-121 on the CRLM dataset, and the F1-score of DenseNet-121 on the RCC dataset. Reporting statistical significance solely for accuracy — without corresponding significance measures for other relevant metrics such as precision or recall — provides an incomplete assessment of model performance. This is particularly concerning in imbalanced classification tasks, where improvements in accuracy may mask critical deficiencies, such as an increase in false positives or false negatives. To ensure scientific rigor and transparency, all key performance metrics reported in the tables — especially those being used to support claims of improvement — should be accompanied by appropriate statistical tests and corresponding p-values. Furthermore, if a reported result is not statistically significant, it should be clearly labeled as such or omitted to avoid misinterpretation. Line 646: “FLOPs of all these models significantly decrease”, the p-value of the significance, the test name, and the sample size for the test are missing for this claim. Minor issues: In the supporting material, the dots on the figures need to be transparent and with a narrow margin for each marker. The points are obstructing each other. The test for 1 p-value has been reported as signrank test but others are missing the statistical test name. Adding a sentence to clarify that the same test has been used to calculate all significance levels could help clarify. Please also include number of samples used to calculate p value (probably k=6). Line 542-546, Lines 542–546 — The term “training crash” is not clearly defined or quantified, making it difficult to interpret its impact or relevance. I recommend moving lines 542–554 to the Discussion section, as this content is largely descriptive and does not report any measurable or validated performance metrics. Relocating this section would improve the logical structure of the Results section by keeping it focused on objective findings. Line 547- Line 547 — The use of the term “excellent” to describe the model’s performance is not appropriate in the context of a scientific report. Descriptive terms such as this are subjective and should be replaced with objective, metric-based statements that accurately reflect the quantitative results. Line 580, Thie seems to be a typo. Line 433 Using the term Experimental is not appropriate for this section. Similarly line 489. The model fitting and testing has been performed in silico and no experiment was performed in a lab. In figure 4, the dashed lines are reporting accuracy over “val”. However, the term “val” needs to be replaced with “test”, as no k-fold validation has been used for the ablation study, and the performance is measured over test set. Similarly, the term “validation performance” needs to be corrected as “performance over the test set”. Line 644, abbreviation of term FLOP has been repeated twice. Reviewer #3: I would like to thank the authors for their careful and comprehensive revisions in response to my previous comments. All of the concerns I raised have been thoroughly addressed, and the manuscript has been substantially improved in both clarity and scientific rigor. I have two additional recommendations that, while not essential, may further enhance the quality and readability of the manuscript: 1. Inclusion of Supplementary Figures (A to C): The t-SNE visualizations currently provided in the Supplementary Information (Figures A to C) offer valuable insights into the intrinsic structure and difficulty of the classification tasks associated with each dataset. If feasible, I recommend incorporating these figures into the main manuscript to strengthen the presentation of the dataset characteristics and facilitate the reader’s understanding of the challenges involved. 2. Presentation of Bootstrap Confidence Intervals: Supplementary Table S3, which reports 95% bootstrap confidence intervals for the mean performance metrics, provides important information on the robustness of the results. I encourage the authors to consider including these intervals directly in the main performance tables of the manuscript (e.g., Tables 2 to 6), so that the uncertainty associated with each metric is readily visible to the reader. These are minor suggestions aimed at improving the accessibility and completeness of the manuscript. I commend the authors for their significant efforts and support the publication of the revised version, with or without the incorporation of these final recommendations. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean? ). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy . Reviewer #1: Yes: Raquel Cumeras Reviewer #2: Yes: Amir Akbarian Reviewer #3: No ********** [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org . Please note that Supporting Information files do not need this step. |
| Revision 2 |
|
MSMCE: A Novel Representation Module for Classification of Raw Mass Spectrometry Data PONE-D-25-08088R2 Dear Dr. Xiong, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Hirenkumar Kantilal Mewada Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #2: All comments have been addressed Reviewer #3: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #2: Yes Reviewer #3: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #2: Yes Reviewer #3: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #2: Yes Reviewer #3: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #2: Yes Reviewer #3: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #2: All my previous concerns have been thoroughly resolved. The revised manuscript is significantly strengthened and presents a clear, well-supported, and insightful contribution to the field. I commend the authors for their rigorous approach, and the depth of their analysis. Congratulations on producing such a strong and impactful paper. Reviewer #3: (No Response) ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean? ). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy . Reviewer #2: Yes: Amir Akbarian Reviewer #3: No ********** |
| Formally Accepted |
|
PONE-D-25-08088R2 PLOS ONE Dear Dr. Xiong, I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team. At this stage, our production department will prepare your paper for publication. This includes ensuring the following: * All references, tables, and figures are properly cited * All relevant supporting information is included in the manuscript submission, * There are no issues that prevent the paper from being properly typeset You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps. Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. You will receive an invoice from PLOS for your publication fee after your manuscript has reached the completed accept phase. If you receive an email requesting payment before acceptance or for any other service, this may be a phishing scheme. Learn how to identify phishing emails and protect your accounts at https://explore.plos.org/phishing. If we can help with anything else, please email us at customercare@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Hirenkumar Kantilal Mewada Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .