Peer Review History
| Original SubmissionJuly 11, 2025 |
|---|
|
Dear Dr. Branchaw, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. ============================== Overall, this work generates a useful framework that can be used to support professional development of researchers. The reviewers suggested the authors clarify the results and some of the figures prior to publication. ============================== Please submit your revised manuscript by Sep 21 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.
If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols . We look forward to receiving your revised manuscript. Kind regards, Amy Prunuske Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information. 3. If the reviewer comments include a recommendation to cite specific previously published works, please review and evaluate these publications to determine whether they are relevant and should be cited. There is no requirement to cite these works unless the editor has indicated otherwise. 4. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice. Additional Editor Comments: The reviewers were enthusiastic about the creation of the Comprehensive Research Development Framework, and noted several areas for improvement. I encourage the authors to address the reviewers' comments regarding the results section and Figure 4. The reviewers suggest that it might be helpful to include a flow chart and that the authors should consider being more explicit about generating a tool that demonstrated consensus across disciplines. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? Reviewer #1: Partly Reviewer #2: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? -->?> Reviewer #1: N/A Reviewer #2: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available??> The PLOS Data policy Reviewer #1: Yes Reviewer #2: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English??> Reviewer #1: No Reviewer #2: Yes ********** Reviewer #1: The manuscript by Branchaw, Butz, & Ayoob aims to develop a comprehensive framework to guide research training, primarily at the graduate and postdoctoral level. This is accomplished first by a literature review to establish a preliminary list of research competencies and then refined through an online survey and two rounds of card-sorting exercises by experienced researchers to organize the competencies into overall categories. The result is eight categories of research skills that contain 57 specific skills/competencies. This is an interesting study. The methodology is well thought out and the resulting framework they’ve developed has genuine applications for graduate student training across disciplines (and likely for post-docs as well). Unfortunately, the manuscript is seriously hobbled by a lack of clarity in the writing of the Results section. I’ve attempted to list specific problems below, but it might be worth the authors getting together to re-outline this section and then use that outline as a basis for a more readable Results section. Lines 69-70: This is an important statement. Is there a citation? It might be worthwhile for the authors to have a flow chart-like figure that explains their process of collecting competencies from the literature > surveying researchers > creating a refined list > open card sorting > creating categories > closed card sorting > refining the framework. Regarding the card-sorting exercises. Is this an evidence-based approach? If so, what are the citations sorting the use of the open and closed sorting approaches. It feels like the Results are missing a section about their literature review. Lines 409-412 - Information about the 79 outcomes prepared for the survey and the 75 that actually appeared in the survey due to a survey error belongs in the Methods section. For the results section, the authors should present the data as though there were a list of 75 outcomes since that's what there is actually data on. Figure 4 - It is not clear what figure 4 is actually supposed to show us or if it is even that important. Is this meant to communicate the % of the 75 learning outcomes judged important by the survey (segregated by discipline)? Is this really something that is important to show? What outcomes were deemed important is certainly interesting and is presented in Table 4. Figure 4's message seems to be "there were things that researchers thought were important" and that's it. Consider omitting (or explaining it better). Table 4 - Perhaps the title should be "Learning Outcomes ... by 80% or More of Overall Respondents". Also, what is the significance of the non-bolded numbers? Just that they are below 80%? Lines 430-432: Are the details about what categories the relatively low scored (variable importance) outcomes really needed here. It distracts from the subsequent report about the higher scoring 50 outcomes. Lines 434-440: I'm completely confused by this section. "Some variation" is reported by discipline but not what those differences are or whether they were statistically different. Then ratings were combined to minimize those differences. If there really are differences between disciplines, then it seems like something worth talking about and maybe analyzing with distribution stats. If not, then just talk about the combined data. Also, the last sentence presents a series of numbers in parentheses. Are these the learning outcomes in Table 1? If so, the format does not correspond (there are no 3.01 or 4.04 in Table 1). Lines 454-455: Again, why this discussion on what wasn't highly scored. Lines 501-502: Shouldn't the authors refer to Table 5? Also, I'm not sure that Tables 5 & 6 are in the right order. Or Table 5 is confusing. The process of going to phase 1 to phase 2 to the final 8 set of categories is shown in table 5, but the process of getting to the last set of 8 categories is based on data in Table 6. Either this needs to be explained better in the text or the order of the tables should be changed. Lines 523-526: I'm not sure I understand this approach of dealing with the learning outcomes that showed weak consensus. It sounds arbitrary, but I suspect it was not. Two potential reasons for lack of consensus were that there were discipline-specific differences in which researchers thought these outcomes should be placed. Another is that the outcome could be placed in multiple categories because it was interpreted differently by different groups. If the latter, then perhaps the right approach by the authors would be to "split" the outcome into two outcomes with each reflecting the category that researchers matched them to. An interesting follow-up experiment would be to repeat the forced choice supporting and see if the split/renamed categories got sorted as the authors predicted/assumed. Lines 549-557: How were the Table 7 data generated. I didn't see anything in the Methods that might explain it. Also, what exactly is it saying? Is this mapping how often a given outcome was discussed in the literature reviewed? Finally, doesn't this belong in the Results section? Line 563 paragraph: It might be worth stating that the CDRF is meant to represent a relative consensus across disciplines. Line 576 paragraph: The use of the CRDF to increase transparency is an important point. Could the authors discuss how to operationalize that (similar to how they discussed using the CRDF as an assessment tool in the subsequent section). Placed in a student handbook? Part of a program/mentor/student compact? Something else? This is somewhat discussed in the limitations section, but regarding the data from the Arts & Humanities. How confident are the authors regarding this data given the small sample size? Perhaps these 5 are in a post-grad research career because they had this research training in their undergrad years. Reviewer #2: “The Comprehensive Research Development Framework (CRDF): Core Learning Outcomes for Research Training” manuscript presents an important framework that the authors developed. The process for developing the framework was excellent – supplementing the authors’ deep expertise with input from multiple stakeholders in multiple ways. The framework will be useful to a range of stakeholders (e.g., program administrators, faculty members, students, postdoctoral scholars) across a wide range of fields – including this reviewer. The manuscript is well written and easy to follow. Below are items that could be improved in the manuscript. -Consistency or simplification of the headings would be helpful. For example, the methods sections has STEPS 1-4, with STEP 1 having four lettered subsections; the results section has two main subsections and one of those subsections lists the two phases of card sorting; and limitations, conclusions, and future directions are all short main sections. -Somewhere in the manuscript it would be helpful to note the geographic comprehensiveness of the framework. For example, were most of the frameworks used in the literature review from the United States (or from the US and Europe) – or was there global representation from where the frameworks were developed? Is researcher development often done similarly globally, or is this framework likely most applicable to US researcher development? -There are some items that could be clarified on the figures and tables. *More detail in the captions would be helpful so that the tables and figures could stand alone. *There are some superscripts in Table 6, and the descriptions of what those superscripts represent is missing (i.e., there are not footnotes below Table 6). *Table 7 has a column titled “framework names,” but the items listed seem to be the article titles. *Table 1 lists the learning outcomes as 1.1, 1.2, 1.3…, and the text of the manuscript adds a “0” before the learning outcomes below 10 (e.g., 3.01 on line 439). ********** what does this mean? ). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy Reviewer #1: No Reviewer #2: Yes: Meghann Jarchow ********** [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org |
| Revision 1 |
|
The Comprehensive Researcher Development Framework (CRDF): Core Learning Outcomes for Research Training PONE-D-25-37622R1 Dear Dr. Branchaw, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. For questions related to billing, please contact billing support . If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Amy Prunuske Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: |
| Formally Accepted |
|
PONE-D-25-37622R1 PLOS ONE Dear Dr. Branchaw, I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team. At this stage, our production department will prepare your paper for publication. This includes ensuring the following: * All references, tables, and figures are properly cited * All relevant supporting information is included in the manuscript submission, * There are no issues that prevent the paper from being properly typeset You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps. Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. You will receive an invoice from PLOS for your publication fee after your manuscript has reached the completed accept phase. If you receive an email requesting payment before acceptance or for any other service, this may be a phishing scheme. Learn how to identify phishing emails and protect your accounts at https://explore.plos.org/phishing. If we can help with anything else, please email us at customercare@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Amy Prunuske Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .