Peer Review History
| Original SubmissionMarch 10, 2022 |
|---|
|
PONE-D-22-07182Is image-to-image translation the panacea for multimodal image registration? A comparative studyPLOS ONE Dear Dr. Lindblad, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. In particular, both reviewers raised the need to cite additional references, as well as concerns about how the some of the evaluation experiments were conducted. Please submit your revised manuscript by Aug 20 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Dzung Pham Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. Thank you for stating the following in the Acknowledgments Section of your manuscript: The authors are financially supported by the Wallenberg AI, 906 Autonomous Systems and Software Program (WASP) AI-Math initiative, VINNOVA 907 (MedTech4Health project 2017-02447) and the Swedish Research Council (project 908 2017-04385). However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows: The authors are financially supported by the Wallenberg AI, Autonomous Systems and Software Program (WASP) AI-Math initiative, VINNOVA (MedTech4Health, project 2017-02447) and the Swedish Research Council (project 2017-04385). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. Please include your amended statements within your cover letter; we will change the online submission form on your behalf. 3. We note that Figures 2 and 4 in your submission contain satellite images which may be copyrighted. All PLOS content is published under the Creative Commons Attribution License (CC BY 4.0), which means that the manuscript, images, and Supporting Information files will be freely available online, and any third party is permitted to access, download, copy, distribute, and use these materials in any way, even commercially, with proper attribution. For these reasons, we cannot publish previously copyrighted maps or satellite images created using proprietary data, such as Google software (Google Maps, Street View, and Earth). For more information, see our copyright guidelines: http://journals.plos.org/plosone/s/licenses-and-copyright. We require you to either (1) present written permission from the copyright holder to publish these figures specifically under the CC BY 4.0 license, or (2) remove the figures from your submission: a. You may seek permission from the original copyright holder of Figures 2 and 4 to publish the content specifically under the CC BY 4.0 license. We recommend that you contact the original copyright holder with the Content Permission Form (http://journals.plos.org/plosone/s/file?id=7c09/content-permission-form.pdf) and the following text: “I request permission for the open-access journal PLOS ONE to publish XXX under the Creative Commons Attribution License (CCAL) CC BY 4.0 (http://creativecommons.org/licenses/by/4.0/). Please be aware that this license allows unrestricted use and distribution, even commercially, by third parties. Please reply and provide explicit written permission to publish XXX under a CC BY license and complete the attached form.” Please upload the completed Content Permission Form or other proof of granted permissions as an "Other" file with your submission. In the figure caption of the copyrighted figure, please include the following text: “Reprinted from [ref] under a CC BY license, with permission from [name of publisher], original copyright [original copyright year].” b. If you are unable to obtain permission from the original copyright holder to publish these figures under the CC BY 4.0 license or if the copyright holder’s requirements are incompatible with the CC BY 4.0 license, please either i) remove the figure or ii) supply a replacement figure that complies with the CC BY 4.0 license. Please check copyright information on all replacement figures and update the figure caption with source information. If applicable, please specify in the figure caption text when a figure is similar but not identical to the original image and is therefore for illustrative purposes only. The following resources for replacing copyrighted map figures may be helpful: USGS National Map Viewer (public domain): http://viewer.nationalmap.gov/viewer/ The Gateway to Astronaut Photography of Earth (public domain): http://eol.jsc.nasa.gov/sseop/clickmap/ Maps at the CIA (public domain): https://www.cia.gov/library/publications/the-world-factbook/index.html and https://www.cia.gov/library/publications/cia-maps-publications/index.html NASA Earth Observatory (public domain): http://earthobservatory.nasa.gov/ Landsat: http://landsat.visibleearth.nasa.gov/ USGS EROS (Earth Resources Observatory and Science (EROS) Center) (public domain): http://eros.usgs.gov/# Natural Earth (public domain): http://www.naturalearthdata.com/ [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The authors present an empirical study evaluating the efficacy of image translations (I2I) techniques for the purpose of multimodal registration. The study evaluates 4 recent deep learning based I2I techniques which are used with mono-modal registrations (SIFT+RANSAC and a-AMD) , and compared against four standard multi-modal methods as benchmarks. The methods were evaluated over 4 distinct biomedical imaging datasets, using Frechet Inception Distance (FID) to evaluate the image translation, and displacement accuracy to evaluation the registration. Overall, I am impressed by this work. The authors were thorough in their analysis and explored a broad sample of techniques. The paper was also well written and reads very clearly. However, I have several concerns/questions I hope the authors can address: 1.) The introduction/background seems to be missing a discussion of image translation-based registration techniques that pre-dates the advent of deep learning in the field. Back then this was often referred to as image synthesis. Here are a few examples of these techniques, although many more were proposed in the literature: Bogovic, J.A., Hanslovsky, P., Wong, A. and Saalfeld, S., 2016, April. Robust registration of calcium images by learned contrast synthesis. In 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI) (pp. 1123-1126). IEEE. Chen, M., Carass, A., Jog, A., Lee, J., Roy, S. and Prince, J.L., 2017. Cross contrast multi-channel image registration using image synthesis for MR brain images. Medical image analysis, 36, pp.2-14. Cao, X., Yang, J., Gao, Y., Wang, Q. and Shen, D., 2018. Region-adaptive deformable registration of CT/MRI pelvic images via learning-based image synthesis. IEEE Transactions on Image Processing, 27(7), pp.3500-3512. 2.) One topic that is rarely discussed in this area, which I hope the authors can touch upon, is the stochastic nature of deep learning based I2I techniques. That is, given the exact same train dataset, it is difficult to rebuild a model to provide the exact same image translation. While this problem isn’t unique to I2I models, it seems to have less of an impact on segmentation or classification networks where the model and results seem to converge quickly. In our experience, this issue gets amplified when further using the l2l model results as inputs for downstream registration, which seem to be very sensitive to this variance. Has the authors explored this limitation of these techniques? And how much impact do they think it has on their results? Is 3 folds sufficient to cover this variance? 3.) I am somewhat confused by the structuring of the comparisons reported in the experimental results. It seems the logical evaluation here would be to use each registration method with/without the image translation results as inputs, and swapping multimodal cost functions (MIND, MI, etc.) for monomodal ones (SSD, NCC, etc.) That way we can see the direct improvement that translating the modality has on the registration. However, this was only done for a-AMD and SIFT, and not fully for SIFT, since it wasn’t presented with a multimodal equivalent. The other methods were compared on just the raw images. This makes it difficult to tell if the loss/gains between the methods were due to the image translation, or differences between the registration technique themselves. For example, we know that Elastix is heavily tuned to perform well on MR images, that would explain the 100% registration accuracy on the MR dataset. However, that high performance is not necessarily because it used MI as a metric, there’s a good chance there are other heuristics built into the registration model that greatly help with the alignment. A more informative test would be to switch Elastix to using SSD instead of MI and use I2I results as inputs and see what happens. (And likewise for the other methods.) 4.) The authors mention that in Table 1, we can see some asymmetry depending on the translation direction. How do we know this difference is due to asymmetric performance of the image translation, and not a property of the different modalities? I.e. Is it possible the FID measures is just higher/lower in general for one modality over the other? Or is it normalized somehow to each modality? 5) In the conclusion, the authors make some strong statements about the efficacy of I2I based registration approaches relative to standard approaches (such as MI) based on their results. However, I think it is important to note that in this work, only intra-subject, rigid alignment, using simulated displacements were considered. This is a fairly limited registration task, and is not a great representation of the majority of real-world registration problems. Other studies have shown that once we start working with more complex transformations, cross subject/specimen data, or temporal shifts in the image, the effectiveness of traditional multi-modal measures such as MI starts falling dramatically, which is the driving motivation for many of the techniques cited by the authors. Reviewer #2: This work tackles the well-known problem of multimodal registration, which lacks proper definitions for similarity functions between aligned images. Recently, due to the advent of many image-to-image translation methods, several methods propose to cast this multimodal registration problem as a monomodal problem, where several similarity functions have proved successful, by means of synthesising source domain images following the appearance of the target domain images. The authors provide a nice trade-off between some of the standard metrics designed for multimodal registration and some methods using the intermediate I2I step with a monomodal similarity metric. Moreover, they add their recently proposed methods CoMIR, which in my opinion could be though also as an I2I but to a latent domain between the source and target domains. The methods, datasets and experiments are well presented in the manuscript, even though sometimes it may be confusing due to the wealth of results and comparisons. I highlight the fact that the methods have been tested in 4 different medical imaging datasets. However, I have some major comments and some minor comments that can be found in the attached PDF. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: Yes: Adrià Casamitjana ********** [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.
|
| Revision 1 |
|
Is image-to-image translation the panacea for multimodal image registration? A comparative study PONE-D-22-07182R1 Dear Dr. Lindblad, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Dzung Pham Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewer 2 suggested that a sentence from Section 5.1 that was removed in the revision be put back. I leave it up to the authors whether they would like to add this sentence back in proofs. Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed Reviewer #2: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The authors have adequately addressed each of my major concerns regarding their work. I have no additional comments. Reviewer #2: Dear authors, Thanks for carefully addressing all concerns from both reviewers and the editors. I think the manuscript has improved in readability and completeness. I mostly agree with all reviewer’s responses, and thanks for justifying your experimental decisions. I just want to add something related to the paired vs. unpaired discussion we began. 1.- Regarding your paragraph in the letter of response: “Being unsupervised I2I translation methods, DRIT++, StarGANv2, and CycleGAN do not benefit from aligned pairs during training (these methods do not use that information even though they are presented with it). On the other hand, pix2pix and CoMIR need to utilise the alignment information for training and cannot be trained without. The qualitative advantage of the “unsupervised” I2I approaches is reasonable, while, as observed in our results, the quantitative benefits of the “supervised” ones are clearly shown.” I completely agree that some methods (e.g., pix2pix and CoMIR) need paired training data by design and that others, such as CycleGAN, DRIT++ and StarGANv2 need not. However, the optimization of the latter seems easier when they are fed with paired images -- even though I don’t know if that is the case in practice (maybe it gets to a local minima that do not generalize well to unseen images). In any case I would keep the sentence where you explain how you sample the images from both modalities depending on the method (Sec. 5.1). Apologies if my comment was confusing in the first place. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: Yes: Adrià Casamitjana Díaz ********** |
| Formally Accepted |
|
PONE-D-22-07182R1 Is image-to-image translation the panacea for multimodal image registration? A comparative study Dear Dr. Lindblad: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr Dzung Pham Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .