Peer Review History

Original SubmissionSeptember 15, 2024
Decision Letter - Jinran Wu, Editor

PONE-D-24-40356Accuracy of lastest large language models in answering multiple choice questions in dentistry: a comparative studyPLOS ONE

Dear Dr. Nguyen,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jan 15 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.
  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.
  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Jinran Wu, PhD

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This is an important and timely study, addressing the evolving role of large language models (LLMs) in dental education and their comparative performance. While the manuscript offers valuable insights, there are areas that require further clarification and expansion to enhance the study's rigor and relevance.

The introduction is clear and contextualizes the importance of LLMs in dental education. However, the rationale for selecting the six specific LLMs should be more detailed. Were they chosen based on popularity, accessibility, or specific capabilities in handling dental questions? This will strengthen the justification for their inclusion. The claim that newer versions perform better than older ones is supported by references, but it would benefit from a brief discussion of the limitations of previous studies (e.g., exclusion of image-based questions).

The study design is robust, adhering to STROBE guidelines, but a few points require elaboration. The selection of MCQs from U.S. board review books is reasonable; however, explain why these books are considered representative of global standards. Were efforts made to ensure that the questions are balanced across sub-disciplines and difficulty levels? The study acknowledges the limited number of image-based questions (47). Discuss whether this sample size is sufficient for meaningful comparisons and how future research might address this limitation. While the use of χ² tests and Bonferroni corrections is appropriate, clarify why these methods were chosen over others, such as logistic regression, which might better account for variations across question types and LLM capabilities.

The results are well-organized and supported by comprehensive tables. To improve clarity and impact, expand on the reasons for variation in LLM performance across disciplines (e.g., why dental anatomy and occlusion questions exhibit lower accuracy). Could this be linked to the spatial reasoning required or the training data limitations of LLMs? Discuss why ChatGPT, Claude, and Copilot had similarly low accuracies for image-based questions. Was this due to limitations in image interpretation algorithms, training data, or question complexity? Figure 1 illustrating Gemini's response to image-based questions is helpful, but additional visuals (e.g., a confusion matrix or graphical comparison of discipline-specific accuracies) would enhance understanding.

The discussion effectively interprets results but lacks depth in some areas. Provide a deeper analysis of why Copilot outperformed ChatGPT despite both being based on GPT architecture. This could include differences in fine-tuning, training data, or optimization for specific tasks. Highlight practical implications of the study for dental educators. For example, how can LLMs be integrated into curricula without compromising the rigor of assessment? The point about the potential misuse of LLMs in examinations is valid but could be expanded. Discuss mitigation strategies, such as proctoring technologies or modified examination formats. While the limitations section is clear, it could include the need to evaluate the models' ability to integrate multimodal data (e.g., combining text and images), potential biases in LLM training data that may influence performance on specialized topics, and a suggestion to incorporate questions from non-U.S. sources to assess generalizability.

The conclusion succinctly summarizes the findings but could emphasize the need for continuous updates to LLMs to handle complex and specialized questions effectively and the importance of balancing LLM reliance with critical thinking skills in dental education and practice.

Minor revisions include ensuring consistency in formatting tables and figures (e.g., uniform font sizes and styles). While the language is clear, minor grammatical edits are needed to improve readability (e.g., "Copilot's accuracy showed a significant improvement" could be rephrased for conciseness). Some references (e.g., those cited for image-based question accuracy) are compelling but could be updated to include the most recent studies in medical and dental education.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Revision 1

December 12th, 2024

Dear Editorial Board, PLOS ONE,

We have revised the manuscript thoroughly according to the comments of the reviewers. Any revisions made in our manuscript document were highlighted in red. Please help us review the manuscript again.

Sincerely,

---------------------------------

Reviewer #1: This is an important and timely study, addressing the evolving role of large language models (LLMs) in dental education and their comparative performance. While the manuscript offers valuable insights, there are areas that require further clarification and expansion to enhance the study's rigor and relevance.

The introduction is clear and contextualizes the importance of LLMs in dental education. However, the rationale for selecting the six specific LLMs should be more detailed. Were they chosen based on popularity, accessibility, or specific capabilities in handling dental questions? This will strengthen the justification for their inclusion. The claim that newer versions perform better than older ones is supported by references, but it would benefit from a brief discussion of the limitations of previous studies (e.g., exclusion of image-based questions).

Thank you for your comment. We have added the criteria for selecting the six LLMs and previous studies' limitations to the introduction section:

However, previous studies on the accuracy of LLMs were performed on the older versions, which lacked the advanced multimodal capabilities now available. Furthermore, these studies excluded image-based questions because older LLM versions could not process image attachments in prompts [8,10]. These critical limitations warrant further research exploring the full potential of current models.

Model selection would be based on popularity, recency, multimodal capabilities, prominence in AI research and applications, accessibility, and ability to address domain-specific questions.

The study design is robust, adhering to STROBE guidelines, but a few points require elaboration. The selection of MCQs from U.S. board review books is reasonable; however, explain why these books are considered representative of global standards. Were efforts made to ensure that the questions are balanced across sub-disciplines and difficulty levels?

Thank you for your comment. We have added the explanation for the use of U.S. Board Review Books as global standards:

All 1490 MCQs available within two board review books for the United States (US) National Board Dental Examination were selected as the sampling base for the study [14,15]. These books are widely recognized for their comprehensive coverage of the dental curriculum and alignment with the national licensing exam. The US boasts numerous top-ranked dental schools, and its rigorous accreditation standards influence dental education globally [8]. Furthermore, these specific review books have an average rating of 4.7 out of 5 stars from 114 global ratings on Amazon, suggesting their acceptance as representative of global standards [16]. While the inclusion of all questions in the books captures their comprehensive nature, it also means that the number of questions varies across dental sub-disciplines, reflecting the emphasis placed on different topics within the books and potentially within the National Board Dental Examination itself.

The study acknowledges the limited number of image-based questions (47). Discuss whether this sample size is sufficient for meaningful comparisons and how future research might address this limitation.

Thank you for your comment. We have added more discussion about the limited number of image-based questions:

First, the number of image-based questions available within the selected resources was limited, lower than the calculated sample size for ideal comparisons. Despite this limitation, statistically significant differences were observed between certain LLMs on these questions. However, future research should focus on the ability to answer image-based questions, including radiographs and histopathology specimens, to enable more robust and generalizable comparisons of LLM performance in this critical domain.

While the use of χ² tests and Bonferroni corrections is appropriate, clarify why these methods were chosen over others, such as logistic regression, which might better account for variations across question types and LLM capabilities.

Thank you for your comment. We have added the reason for choosing χ² tests and Bonferroni corrections:

While methods like logistic regression could potentially account for variations across question types and LLM capabilities, χ² tests with Bonferroni corrections were deemed more appropriate for this study due to their focus on comparing proportions and their robustness to small sample sizes in certain categories. Conversely, logistic regression focuses on predicting the likelihood of a correct answer based on multiple variables.

The results are well-organized and supported by comprehensive tables. To improve clarity and impact, expand on the reasons for variation in LLM performance across disciplines (e.g., why dental anatomy and occlusion questions exhibit lower accuracy). Could this be linked to the spatial reasoning required or the training data limitations of LLMs? Discuss why ChatGPT, Claude, and Copilot had similarly low accuracies for image-based questions. Was this due to limitations in image interpretation algorithms, training data, or question complexity?

Thank you for your suggestion. We have added more discussion about the reason for variation in LLM performance:

Additionally, this discrepancy may stem from gaps in the LLMs' training data, particularly regarding spatial reasoning and visualization skills necessary for questions related to dental anatomy and occlusion.

This lower accuracy may be attributed to limitations in LLM image interpretation algorithms, potentially stemming from their architectural design or insufficient exposure to diverse medical images during training, particularly those related to dentistry. Furthermore, the inherent ambiguity and subjectivity in interpreting certain images, even for human experts, could pose an additional challenge for these LLMs.

Figure 1 illustrating Gemini's response to image-based questions is helpful, but additional visuals (e.g., a confusion matrix or graphical comparison of discipline-specific accuracies) would enhance understanding.

Thank you for your suggestions. We have added Fig. 2 and Fig.3 for better visualization.

Fig. 2. Grouped stacked bar chart presenting the accuracy of large language models for text-based, image-based questions and the total sample. Percentages with the same lowercase letter were not statistically different, as determined by the post-hoc analysis with Bonferroni correction.

Fig. 3. Grouped stacked bar chart presenting the accuracy of large language models for individual dental disciplines. (A) Anatomic Sciences (AS), Biochemistry & Physiology (BP), Microbiology & Pathology (MP), Dental Anatomy & Occlusion (DAO), Pharmacology (PH), Operative Dentistry & Prosthodontics (ODP), Oral & Maxillofacial Surgery (OMF). (B) Orthodontics (OR), Pediatric Dentistry (PD), Endodontics (EN), Periodontics (PE), Radiology (RA), Pathology (PA), Patient Management & Public Health (PMPH).

The discussion effectively interprets results but lacks depth in some areas. Provide a deeper analysis of why Copilot outperformed ChatGPT despite both being based on GPT architecture. This could include differences in fine-tuning, training data, or optimization for specific tasks.

Thank you for your comment. We have provided a deeper analysis according to your comment:

However, it is also important to consider that variations in training data and fine-tuning strategies can still impact the performance of LLMs even when they share the same underlying architecture [17]. Interestingly, despite Copilot’s fine-tuning and optimization in code generation [18], it performed slightly better in both dental text-based and image-based questions compared to ChatGPT. Variations in training data, such as the inclusion of domain-specific datasets or the emphasis on technical accuracy, may further contribute to Copilot’s edge in handling specialized questions.

Highlight practical implications of the study for dental educators. For example, how can LLMs be integrated into curricula without compromising the rigor of assessment? The point about the potential misuse of LLMs in examinations is valid but could be expanded. Discuss mitigation strategies, such as proctoring technologies or modified examination formats.

Thank you for your suggestion. We have added the practical implications for the study according to your comments:

These models could be integrated into curricula as supplemental aids for self-directed learning, offering instant explanations and additional resources. However, to ensure the rigor of assessments is not compromised, educators should use LLMs strategically by focusing on enhancing students’ critical thinking and problem-solving skills rather than relying solely on AI-generated responses. On the other hand, traditional assessment methods for evaluating core competencies in clinical decision-making should be maintained.

To mitigate these risks, institutions can adopt strategies such as advanced proctoring technologies that monitor for unauthorized devices or altered examination formats that emphasize critical thinking and problem-solving over rote memorization.

While the limitations section is clear, it could include the need to evaluate the models' ability to integrate multimodal data (e.g., combining text and images), potential biases in LLM training data that may influence performance on specialized topics, and a suggestion to incorporate questions from non-U.S. sources to assess generalizability.

Thank you for your comment. We have add more discussion on these limitations to the study limitation:

Third, the ability of LLMs to gather patient information and integrate multimodal data, such as combining text and images for comprehensive problem-solving, was not evaluated. Fourth, potential biases in the training data of LLMs may influence their performance on specialized topics. Furthermore, the study's reliance on US-based questions may limit generalizability. Future research should include diverse, non-US sources to assess the models' robustness across different educational contexts.

The conclusion succinctly summarizes the findings but could emphasize the need for continuous updates to LLMs to handle complex and specialized questions effectively and the importance of balancing LLM reliance with critical thinking skills in dental education and practice.

Thank you for your comment. We have modified the conclusions according to your comment:

However, accuracy on image-based questions remains low, around 60%, underscoring the need for continuous updates to LLMs to better handle complex and specialized questions. LLMs capable of addressing both text and image-based queries outperformed those limited to text alone. Dental clinicians and students should prioritize using the most up-to-date LLMs while also balancing their reliance on these tools with the development of critical thinking skills to ensure sound decision-making in learning, clinical practice, and research.

Minor revisions include ensuring consistency in formatting tables and figures (e.g., uniform font sizes and styles).

Thank you for your comment. The table's large size required a smaller font to ensure readability within the document. We will ensure consistency in formatting, including uniform font sizes and styles, during the final typesetting process.

While the language is clear, minor grammatical edits are needed to improve readability (e.g., "Copilot's accuracy showed a significant improvement" could be rephrased for conciseness).

Thank you for your comment. We have carefully reviewed the manuscript and made the necessary grammatical edits to enhance readability and clarity.

Copilot's accuracy improved significantly from 72.6% in the previous study to 85.5% in this study [10].

Some references (e.g., those cited for image-based question accuracy) are compelling but could be updated to include the most recent studies in medical and dental education.

Thank you for your comment. We have added more recent references to ensure the citations are comprehensive and reflect the latest advancements in the field.

Attachments
Attachment
Submitted filename: Response to reviewer.docx
Decision Letter - Jinran Wu, Editor

PONE-D-24-40356R1Accuracy of lastest large language models in answering multiple choice questions in dentistry: a comparative studyPLOS ONE

Dear Dr. Nguyen,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Feb 10 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.
  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.
  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Jinran Wu, PhD

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: (No Response)

********** 

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

********** 

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: N/A

Reviewer #2: Yes

********** 

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

********** 

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

********** 

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors have addressed all my concerns.

Congrats! Before the final acceptance, please polish the English writing.

Reviewer #2: The article aims to evaluate the accuracy of responses provided by current-generation Large Language Models (LLMs) to multiple-choice questions (MCQs) in the field of dentistry. The study is interesting both in terms of its chosen subject and methodological approach. Considering the lack of literature on the potential impact of LLMs in dental education and clinical practice, this work may provide a significant contribution. Methodologically, the performances of different LLMs have been comparatively and statistically analyzed. Overall, the article is well-structured, fluent, and written in a clear language.

I have only two suggestions.

First, the authors emphasize that they always used the most up-to-date LLM. However, given the rapid evolution of LLMs, it should be made clearer that this emphasis applies to August 2024. Currently, it is only mentioned once in the methodology section and not stated in the abstract.

The other suggestion concerns clinical reasoning mentioned within the article. There are studies in which LLMs are not only used to solve questions but also to produce questions utilizing clinical reasoning. At least a brief mention of this point could be included in the introduction regarding its relevance for medical/dental education. There are studies in the literature on LLM prompts designed to measure clinical reasoning skills. A recent study has even generated Script Concordance Test questions, for example:

Kıyak, Y. S., & Emekli, E. (2024). Using Large Language Models to Generate Script Concordance Test in Medical Education: ChatGPT and Claude. Spanish Journal of Medical Education, 6(1). https://doi.org/10.6018/edumed.636331

********** 

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Revision 2

December 12th, 2024

Dear Editorial Board, PLOS ONE,

We have revised the manuscript thoroughly according to the comments of the reviewers. Any revisions made in our manuscript document were highlighted in red. Please help us review the manuscript again.

Sincerely,

---------------------------------

Reviewer #1: The authors have addressed all my concerns.

Congrats! Before the final acceptance, please polish the English writing.

Thank you for your commendation. We have carefully reviewed the manuscript for grammar and language errors.

--------------------------------

Reviewer #2: The article aims to evaluate the accuracy of responses provided by current-generation Large Language Models (LLMs) to multiple-choice questions (MCQs) in the field of dentistry. The study is interesting both in terms of its chosen subject and methodological approach. Considering the lack of literature on the potential impact of LLMs in dental education and clinical practice, this work may provide a significant contribution. Methodologically, the performances of different LLMs have been comparatively and statistically analyzed. Overall, the article is well-structured, fluent, and written in a clear language.

I have only two suggestions.

First, the authors emphasize that they always used the most up-to-date LLM. However, given the rapid evolution of LLMs, it should be made clearer that this emphasis applies to August 2024. Currently, it is only mentioned once in the methodology section and not stated in the abstract.

Thank you for your comment. We have added this clarification to the abstract:

This study evaluated six of the latest LLMs as of August 2024, including ChatGPT 4.0 omni (OpenAI), Gemini Advanced 1.5 Pro (Google), Copilot Pro with GPT-4 Turbo (Microsoft), Claude 3.5 Sonnet (Anthropic), Mistral Large 2 (Mistral AI), and Llama 3.1 405b (Meta).

The other suggestion concerns clinical reasoning mentioned within the article. There are studies in which LLMs are not only used to solve questions but also to produce questions utilizing clinical reasoning. At least a brief mention of this point could be included in the introduction regarding its relevance for medical/dental education. There are studies in the literature on LLM prompts designed to measure clinical reasoning skills. A recent study has even generated Script Concordance Test questions, for example:

Kıyak, Y. S., & Emekli, E. (2024). Using Large Language Models to Generate Script Concordance Test in Medical Education: ChatGPT and Claude. Spanish Journal of Medical Education, 6(1). https://doi.org/10.6018/edumed.636331

Thank you for your suggestion. We have revised the introduction section and added your suggested reference:

Beyond their use in answering clinical questions, LLMs are also being explored for their potential to generate questions that assess clinical reasoning skills, a crucial aspect of medical and dental education [4].

Attachments
Attachment
Submitted filename: Response to reviewer R2.docx
Decision Letter - Jinran Wu, Editor

Accuracy of lastest large language models in answering multiple choice questions in dentistry: a comparative study

PONE-D-24-40356R2

Dear Dr. Nguyen,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Jinran Wu, PhD

Academic Editor

PLOS ONE

Formally Accepted
Acceptance Letter - Jinran Wu, Editor

PONE-D-24-40356R2

PLOS ONE

Dear Dr. Nguyen,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Jinran Wu

Academic Editor

PLOS ONE

Open letter on the publication of peer review reports

PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.

We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.

Learn more at ASAPbio .