Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Physician awareness of, interest in, and current use of artificial intelligence large language model-based virtual assistants

  • Rachel L. Solmonovich ,

    Roles Conceptualization, Data curation, Investigation, Methodology, Writing – original draft, Writing – review & editing

    RSolmonovich@northwell.edu

    Affiliations Northwell, New Hyde Park, New York, United States of America, Department of Obstetrics and Gynecology, South Shore University Hospital, Bay Shore, New York; Zucker School of Medicine at Hofstra/Northwell, Hempstead, NY, United States of America

  • Insaf Kouba,

    Roles Conceptualization, Methodology, Writing – review & editing

    Affiliations Northwell, New Hyde Park, New York, United States of America, Department of Obstetrics and Gynecology, South Shore University Hospital, Bay Shore, New York; Zucker School of Medicine at Hofstra/Northwell, Hempstead, NY, United States of America

  • Ji Y. Lee,

    Roles Formal analysis

    Affiliation Biostatistics Unit, Office of Academic Affairs, Northwell Health, New Hyde Park, New York, United States of America

  • Kristen Demertzis,

    Roles Investigation, Methodology, Writing – review & editing

    Affiliations Northwell, New Hyde Park, New York, United States of America, Department of Obstetrics and Gynecology, South Shore University Hospital, Bay Shore, New York; Zucker School of Medicine at Hofstra/Northwell, Hempstead, NY, United States of America

  • Matthew J. Blitz

    Roles Conceptualization, Methodology, Supervision, Writing – review & editing

    Affiliations Northwell, New Hyde Park, New York, United States of America, Department of Obstetrics and Gynecology, South Shore University Hospital, Bay Shore, New York; Zucker School of Medicine at Hofstra/Northwell, Hempstead, NY, United States of America, Institute of Health Systems Science, Feinstein Institutes for Medical Research, Northwell Health, New Hyde Park, New York, United States of America

Abstract

There is increasing medical interest and research regarding the potential of large language model-based virtual assistants in healthcare. It is important to understand physicians’ interest in implementing these tools into clinical practice, so preceding education could be implemented to ensure appropriate and ethical use. We aimed to assess physician 1) awareness of, 2) interest in, and 3) current use of large language model-based virtual assistants for clinical practice and professional development and determine the specific applications of interest and use. Additionally, we wanted to determine associations with age, gender, and role. We conducted a cross-sectional study between 11/08-12/2023 via an anonymous web-based survey that was disseminated among physicians at a large NY healthcare network using snowball sampling. Descriptive and basic inferential statistics were performed. There were 562 respondents, largely males (55.7%), attending physicians (68.5%), and from nonsurgical specialties (67.4%). Most were aware of large language model chatbots (89.7%) and expressed interest (97.2%). Only a minority incorporated it into their practice (21%). Highest levels of interest were for journal review, patient education, and documentation/dictation (88.1-89.5%). The most frequently employed uses were medical information and education and study/research design. Females showed higher interest than males (99.2% vs. 95.5%, p = 0.011). Attendings were more aware of large language models (92.2% vs. 84.2%, p = 0.004), while trainees had increased rates of use (28.8% vs. 17.4%, p = 0.002). Use varied across age brackets, highest among 20-30 year olds (29.1% vs. 13.5%-23.4%, p = 0.018), except for documentation/dictation, where highest use was among the 41-50 year old group (10.5% vs. 2.6%-8.7%, p = 0.047). We concluded that physicians are interested in large language model-based virtual assistants, a minority are implementing it into their practice, and gender-, role-, and age-based disparities exist. As physicians continue to integrate large language models into their patient care and professional development, there is opportunity for research, education, and guidance to ensure an inclusive, responsible, and safe adoption.

Introduction

ChatGPT (generative pre-trained transformer), an open-access artificial intelligence (AI) deep learning large language model (LLM) released in November 2022, inspired the science community to explore the healthcare potential of LLMs. A Pubmed search for “ChatGPT” yielded more than 3,000 results as of May 2024, with a large amount of data examining the performance of LLMs in passing medical exams or assisting with diagnostics. A number of publications demonstrate how ChatGPT has been found to pass various medical licensing exams, [14] suggesting a potential use for medical education and assistance in consultation, diagnosis, and other aspects of patient care. However, there’s a paucity of data related to physicians’ interest in and current use of these tools.

Theoretical advantages for LLMs in the medical field include improved scientific writing, [5] documentation, [1,6,7] dataset analysis, drafting papers, language review, and personalized learning, [5] offering the potential to enhance medical education, clinical decision-making, research, and patient care. [1,8,9] However, concerns exist regarding the implementation of LLMs into clinical practice include ethical, copyright, transparency, and legal issues, and the risk of bias, plagiarism, misinformation, and incorrect citation. [5,1012] These concerns emphasize the need for oversight, regulations, and boundaries to ensure ethical and transparent usage, as well as quality, relevant, and appropriate results that complement current practice.

Physicians may be unaware that these tools exist as possible aids for their clinical and educational needs, and they may have interest in incorporating them into their practice. Proper education on appropriate use, including awareness of intrinsic biases and limitations and appropriate prompt engineering, should precede incorporation of LLMs into daily practice. Some may already be taking advantage of these services, without fully understanding their potential promise and pitfalls. Understanding physician awareness of, interest in, and current use of LLMs would allow for tailoring implementation strategies and the preceding education necessary for safe, ethical, and appropriate use.

The few studies that have looked at healthcare interest in clinical AI were either limited to trainees, were not specific to physicians, or had a small number of survey respondents. [1316] Given the theoretical advantages for LLMs in the medical field and the limited data on how practicing physicians perceive and utilize these tools in real-world settings, we aimed to assess United States physicians’ awareness of, interest in, and current use of LLM-based virtual assistants. Secondary objectives were to determine interest and current use of specific education, research, and patient care purposes, and to stratify interest in and current use of LLMs by gender (male versus female), age (by decade), and role (attending physician versus trainee).

Materials and methods

This was a cross-sectional survey study distributed between November 8, 2023 and December 5, 2023 via institutional email to physicians employed at a large, academic, New York healthcare system. All physicians were eligible to participate, including attendings, fellows, and residents. The Northwell Health institutional review board approved this study and waived the requirement for informed consent.

An established, validated instrument did not exist to address our objectives, so the 15-question survey (S1 File) was developed by the primary author. ChatGPT was utilized during the initial phase of drafting the survey. The final version used in the study was significantly modified to contain questions relevant to our study objectives that were appropriate for our intended survey recipients in a format allowing for ease of completion and data analysis. The survey was reviewed for clarity and comprehensiveness by multiple board-certified attendings, to ensure that the questions presented would satisfy the research objectives. The authors determined that given the exploratory nature and descriptive goals for the study, this was sufficient to finalize the survey design prior to distribution.

The survey link was disseminated through snowball sampling. The survey was first sent out to institutional email addresses of graduate medical education (GME) and department leadership, requesting that they share it with their colleagues. The survey was closed to further responses after 4 weeks.

Before survey initiation, all participants read a short introduction explaining the survey’s purpose and that participation was voluntary and anonymous. The first set of questions covered demographics, such as role, years in practice, specialty, main healthcare setting, age, and gender. Additional questions addressed knowledge of, interest in, and use of LLM-based virtual assistants for various educational and clinical purposes. Levels of interest were ascertained via a 3-point Likert scale (not interested, somewhat interested, very interested). Study data were collected and managed using Research Electronic Data Capture (REDCap) hosted at Northwell Health. [17, 18]

The primary outcomes were rates of physician 1) awareness of, 2) interest in, and 3) current use of LLM-based chatbots. To assess physician interest as a binary variable, a composite variable was created – “no” when answered”not interested” and “yes” when answered “somewhat interested” or “very interested” for any of the questions addressing interest. The secondary outcomes were 1) interest levels in and 2) current use of specific educational and clinical applications. Sub-analyses were performed to determine whether results were associated with gender, age, and role. Age groups were analyzed by decades for practicality and ease of interpretation and comparison.

For categorical variables, frequencies and percentages were tabulated for the overall population and stratified by gender, age, and role. Additionally, frequencies and percentages of demographic variables were tabulated for respondents who reported current LLM use. A total of 51 univariate analyses were performed to assess the association between demographic factors (gender, age, and role) and the survey responses. For each pair of categorical variables, contingency tables were generated, and associations were evaluated using the Chi-squared test. When the expected counts in any cell were less than 5, Fisher’s Exact test was used instead. All analyses were conducted using SAS V9.4 (SAS Institute Inc., Cary NC.) A p-value < 0.05 was considered statistically significant. The study methodology is delineated in Fig 1.

Results

A total of 562 physicians completed the survey. The majority of respondents self-identified as male (55.7%), attending physician role (68.5%), and nonsurgical specialty (67.4%). Baseline demographics are described in Table 1.

thumbnail
Table 1. Baseline demographics of survey responders (n = 562).

https://doi.org/10.1371/journal.pone.0320749.t001

Physicians from 32 specialties completed the survey. The most represented specialties were obstetrics and gynecology (14.2%), pediatrics (12.0%), general surgery (8.6%), anesthesiology (7.9%), internal medicine (7.4%), and orthopedics (5.2%). A few identified as being from an unlisted specialty (4.1%).

Primary outcomes

Most were aware of LLM-based virtual assistants for medical support (89.7%, n = 504) and were interested in using such assistance (97.2%, n = 546). Approximately 1/5 of respondents were already incorporating LLM assistants into their routines (21.0%, n = 118).

Secondary outcomes

1. Interest.

All specialties had high interest rates (85.71%-100%), shown in Table 2, and interest was expressed for all listed purposes. Highest levels of interest were for journal review (89.5%), patient education (88.1%), and documentation/dictation (88.4%). (Table 3, Fig 2)

thumbnail
Table 3. Physician interest in LLM-based virtual assistants (n = 562).

https://doi.org/10.1371/journal.pone.0320749.t003

thumbnail
Fig 2. Physician interest levels and applications of LLM-based virtual assistants.

Interest was expressed for all listed purposes, with highest levels for journal review, patient education, and documentation/dictation. Among the physicians currently using LLM assistants (n = 118), the most frequently used applications were medical information and education, study/research design, and documentation/dictation. *Interest levels were obtained from the entire study cohort (n = 562).

**Current use rates were obtained from a sub-cohort of physicians who endorsed currently using LLM-based virtual assistants, and multiple applications could be endorsed by the same physician (n = 118).

https://doi.org/10.1371/journal.pone.0320749.g002

2. Current use.

Among the physicians currently using LLM assistants (n = 118), the majority were male (57.8%), attending physicians (56.8%), and from nonsurgical specialties (76.9%). Most use was among the 31-40 years age bracket (30.5%). Among the specialties with the highest response rates, use ranged from 6.82%-46.34%, (anesthesiology and internal medicine, respectively). (Table 2). The most frequently used applications were medical information and education (53.4%), study/research design (38.1%), and documentation/dictation (33.1%). (Table 4)

thumbnail
Table 4. Characteristics of physicians currently using LLM-based virtual assistants (n = 118).

https://doi.org/10.1371/journal.pone.0320749.t004

Sub-analyses (Table 5)

thumbnail
Table 5. Inferential statistics by gender, role, and age bracket.

https://doi.org/10.1371/journal.pone.0320749.t005

Gender.

There was no evidence of a statistically significant association between gender and awareness of LLM-based virtual assistants.

Females were more interested than males (99.2% vs. 95.5%, p = 0.011) in using LLM-based virtual assistants. When evaluating level of interest for specific applications, males were less likely to show interest for the following domains: study/research design (20.6% vs. 11.1%, p = 0.011), journal review (13.8% vs. 6.6%, p = 0.015), patient education (17.0% vs. 5.8%, p < 0.001), and exam preparation (32.1% vs. 13.6%, p < 0.001). (Fig 3) There was no evidence of an association between current use of LLM-based virtual assistants and gender.

thumbnail
Fig 3. Physician Interest Rates by Gender and Role.

Males were less likely to show interest for the following domains: study/research design (p = 0.011), journal review (p = 0.015), patient education (p < 0.001), and exam preparation (p < 0.001). More trainees were very interested for the following uses: medical information and education (p = 0.021), documentation/dictation (p = -0.022), study/research design (p = 0.002), journal review (p = 0.002), and exam preparation (p < 0.001).

https://doi.org/10.1371/journal.pone.0320749.g003

Role.

There was a statistically significant association between role, attending versus trainee, and awareness of LLM-based virtual assistants (92.2% vs. 84.2%, respectively; p = 0.004).

While overall interest in using LLM-based assistants did not significantly differ by role, significant associations were found when analyzing interest levels for specific purposes. Specifically, more trainees were very interested for the following uses: medical information and education (47.5% vs 42.3%, p = 0.021), documentation/dictation (68.8% vs. 60.5%, p = -0.022), study/research design (56.3% vs. 45.8%, p = 0.002), journal review (65.3% vs. 50.7%, p = 0.002), and exam preparation (51.4% vs. 38.2%, p < 0.001). (Fig 3) There was a statistically significant association between role and current use of LLMs, with fewer attendings using LLM assistants compared to trainees (17.4% vs. 28.8%, p = 0.002). Within the specific applications for usage, there were significant associations revealed for the following domains: study and research design (5.2% vs. 14.1%, p < 0.001), journal review (3.1% vs. 10.2%, p < 0.001), patient education (3.4% vs. 10.7%, p < 0.001), and exam preparation (1.8% vs. 5.1%, p = 0.031).

Age.

There was no statistically significant association found between age group and awareness of LLM-based virtual assistants.

All individuals between 20-30 years olds were interested, while those in the other age groups expressed varying rates of interest (95.5%-97.6%, p = 0.167). When evaluating level of interest for specific purposes, there were several statistically significant associations between age group and interest level, including for medical information and education, documentation and dictation, study and research design, journal review, and exam preparation.

There was a statistically significant association between age and current use of LLMs. Among respondents aged 20-30 years, 29.1% currently use LLM-based virtual assistants, while only 20.9% in the 31-40 age group, 23.4% in the 41-50 age group, and 13.5% in the 50 + age group use LLM assistance (p = 0.018). When analyzing the specific domains for use, only medical information and education and case discussions were not significant.

Discussion

Principal findings

Physicians from a wide variety of specialties completed the survey. The majority of participants were aware of and interested in LLM-based virtual assistants for clinical and educational purposes, while only a minority were already incorporating it into their practice. Highest interest levels were observed for journal review, patient education, and documentation and dictation. The most common uses were for medical information and education, study and research design, and documentation and dictation. Sub-analyses revealed significant gender-, role-, and age-based differences in awareness, interest, and current use of LLM-based virtual assistants.

The differences observed in the sub-analyses in interest and use of LLM virtual assistants likely reflect a combination of generational attitudes toward technology, professional dynamics, and different demands and workflows. Trainee physicians typically fall within the younger age range, and the findings from the sub-analyses align closely with the interest and usage rates observed in the youngest cohort of physicians, as expected. Younger physicians displayed higher usage rates, possibly due to their increased exposure to technology during training and greater comfort integrating digital tools into their practice. An exception was the higher use of documentation and dictation tools among 41-50 year olds, which might reflect their position as mid-career professionals balancing high clinical loads with administrative tasks, making such tools particularly valuable. This older age group is likely the last to have trained before the implementation of the Health Information Technology for Economic and Clinical Health Act, requiring them to adapt to new electronic medical record systems. As a result, further integration of LLM assistants may have been a natural progression for them. The higher interest levels among female physicians may originate from a greater recognition of the potential for such tools to enhance efficiency in tasks traditionally associated with administrative burden, which often disproportionally impact female physicians. [19]

Results in the context of what is known

ChatGPT amassed a historic amount of users in a brief period, 1 million users within 5 days and 100 million users in two months. [20] Our study demonstrates both interest and use among physicians across specialties. Studies have shown that LLMs have great potential to support research, because they can explore literature and generate hypotheses, handle complex data and extract useful information, and translate complicated findings into more easily understandable language. [21] In a survey by Banerjee et al., the majority of trainee doctors agreed that AI would have an overall positive impact on their training and education. They were most optimistic that clinical AI would enhance their training by improving the efficiency of research, freeing up time to spend on other educational activities, and keeping up with evidence-based practices. [14] The small survey by Spotnitz concluded that practicing clinicians, roles were not specified, were encouraging of using LLMs, especially in assistive roles, but there is need for human oversight. [16] The most positively rated uses were clinical practice and education tasks. Among our surveyed resident and attending physicians, these were also areas of high interest.

A survey by Temsah et al. of healthcare workers in Saudi Arabia found that most were comfortable incorporating ChatGPT into their practice but expressed concerns about credibility and the sources of provided information. Among their survey respondents, which was inclusive of medical students, nurses, technicians, therapists, pharmacists, and physicians, 18.4% were using ChatGPT, and 84.1% of those who were not using it yet were expecting to in the future. [15] Our results are concordant: physicians are interested in using LLMs for a variety of clinical and educational applications, and a minority of them already do. Chen et al. showed via an international web-based questionnaire that physicians and medical students had a positive but reserved attitude toward the application of clinical AI, but they lacked practical experience. They reported a similar clinical AI utility rate of 20% but a much lower awareness rate (38% vs. 89.7%) than our physicians-only cohort. [13]

Clinical implications

Physicians express high interest levels in integrating LLMs into their practice, and a minority already have. Implementing and expanding LLM use in the domains that physicians expressed interest in, and concordantly spend the bulk of their time, could be time-saving and free up time for other purposes. It could also enhance the physician-patient relationship by improving patient education, [1,6,7] comprehension, and retention. Our survey respondents expressed a high level of interest in using LLM assistants for patient education, and a recent study showed that LLMs have potential to generate proficient medical counseling templates in Spanish for patients [22] as well as generate useful handoff notes to reduce physician documentation burden. [23] These are examples in which LLMs can be applied clinically and impact patient care.

Possible reasons for the discrepancy between awareness and interest versus current use rates include lack of training, institutional hesitancy, technological limitations, and ethical and accuracy concerns. However, with high interest within the healthcare community, it is prudent that the adoption of LLMs in medicine be shaped by medical professionals who can impart the appropriate training data and testing prior to their integration into medical practice. [24] Tailoring implementation strategies based on distinct preferences and needs within roles and age groups, as well as focusing on identified areas of interest, may help ensure higher uptake with appropriate preceding education to ensure safe and ethical use.

It is important to consider the ethics involved with integration of LLMs into clinical practice. Significant ethical concerns exist, such as the possibility of perpetuating biases related to race, sex, language, and culture given that their responses reflect their training data, which originates from high-income English-speaking countries. This also increases the risk for limited perspectives and inability to generalize responses to people from other regions of the world. [12] Furthermore, with AI technology in general, there are privacy and confidentiality concerns related to sharing patient information and images, as well as economic and accessibility disparities that may widen for those who cannot afford or access these AI tools. [25] The epistemic opacity, which is the inability to understand how algorithms make decisions, and whom to assign responsibility when error occurs are additional ethical concerns, [26] which need to be addressed prior to broad integration into medical practice.

AI is an innovative tool with potential to revolutionize modern healthcare practices by enhancing both provider and patient experiences and outcomes. It should be seen as an adjunct resource, like textbooks and other digital tools, and integrated thoughtfully within the broader spectrum of clinical decision-making aids. Explainable AI, which helps clinicians understand the AI’s decision-making process, can support physicians by fostering informed clinical practice, thereby eliminating the need to choose between reliance on opaque algorithms or avoiding these valuable tools. [27] LIME (Local Interpretable Model-agnostic Explanations) and SHAP (SHapley Additive exPlanations) are widely used techniques for explaining machine learning model predictions. LIME approximates model behavior using simpler surrogate models to generate quick, interpretable insights, shedding light on the factors contributing to a particular prediction made by a given machine learning or deep learning model. [28] SHAP explains the importance of each individual feature in the context of a particular prediction, identifying key factors exerting the most substantial influence on a given prediction to reveal the inner workings of machine learning or deep learning models. [29] Although LLMs do not inherently use SHAP or LIME to generate responses, they can be applied externally to explain the outputs of such models when interpretability is required and to enable greater transparency and trust in their applications. Additionally, given that LLMs interact directly with their users, they can clearly convey the reasoning behind their conclusions. By sharing reference sources, LLMs can further enhance transparency and trust, allowing users to validate their claims.

Research implications

Our study found that most physicians are aware of and interested in LLM-based chatbots for healthcare purposes, yet only a minority are already using them. The discrepancies observed between demographic groups suggest that the adoption of LLM-based tools may be influenced by generational factors within medicine, in addition to practical needs, which warrants further exploration.

Future research can confirm these findings and investigate the underlying reasons for the discrepancy between the high levels of awareness and interest and their relatively low current use rates. Additionally, studies should explore causal factors influencing LLM adoption and continue to explore the benefits of incorporating LLMs into physician practice, such as saving time and money, which can be allocated to other purposes, as well as physician professional satisfaction. Future projects could also collect qualitative insights about the perspectives of physicians regarding LLM adoption. With increased uptake of LLMs into physician practice, studies exploring the limitations and ethical implications of LLM-based virtual assistants in healthcare should be performed to provide reassurance. Furthermore, additional investigation can provide insight into variations in trends and opinions among medical specialties and geographical locations, as well as among other healthcare providers, such as mid-level providers and nurses.

Strengths and limitations

Our method of survey dissemination via participant self-selection and snowball sampling, which may promote responses of like-minded individuals, as well as only surveying physicians within a NY healthcare system, may limit the generalizability of our results. Additionally, due to our survey distribution methods, an accurate denominator needed to calculate a response rate could not be determined. However, a strength of our study is the large number of responses from physicians across different practice types, specialties, and with varying levels of experience.

Conclusions

Our study summarizes the current awareness, interest, and adoption of LLM-based virtual assistants by physicians. Physicians express high interest levels in integrating LLMs into their practice, with a notable gap between the amount of interest and actual use rates, both of which are associated with gender, role, and age variations. There is also discordance between which LLM applications physicians are interested in versus what they are currently using. There is still opportunity to implement the necessary research, education, and guidance to ensure an inclusive, beneficial, responsible, and safe adoption of LLMs into physician education and clinical practice.

Supporting information

S1 File. Survey assessing physician interest in and use of AI-powered virtual assistants.

https://doi.org/10.1371/journal.pone.0320749.s001

(PDF)

References

  1. 1. Lee P, Bubeck S, Petro J. Benefits, limits, and risks of gpt-4 as an ai chatbot for medicine. N Engl J Med. 2023;388(13):1233–9. pmid:36988602
  2. 2. Kung TH, Cheatham M, Medenilla A, Sillos C, De Leon L, Elepaño C, et al. Performance of ChatGPT on USMLE: Potential for AI-assisted medical education using large language models. PLOS Digit Health. 2023;2(2):e0000198. pmid:36812645
  3. 3. Oztermeli AD, Oztermeli A. ChatGPT performance in the medical specialty exam: an observational study. Medicine (Baltimore). 2023;102(32):e34673. pmid:37565917
  4. 4. Gilson A, Safranek CW, Huang T, Socrates V, Chi L, Taylor RA, et al. How Does ChatGPT Perform on the United States Medical Licensing Examination (USMLE)? The Implications of Large Language Models for Medical Education and Knowledge Assessment. JMIR Med Educ. 2023;9:e45312. pmid:36753318
  5. 5. Sallam M. ChatGPT utility in healthcare education, research, and practice: systematic review on the promising perspectives and valid concerns. Healthcare (Basel). 2023;11(6):887. pmid:36981544
  6. 6. Decker H, Trang K, Ramirez J, Colley A, Pierce L, Coleman M, et al. Large language model-based chatbot vs surgeon-generated informed consent documentation for common procedures. JAMA Netw Open. 2023;6(10):e2336997. pmid:37812419
  7. 7. Bala S, Keniston A, Burden M. Patient perception of plain-language medical notes generated using artificial intelligence software: Pilot mixed-methods study. JMIR Form Res. 2020;4(6):e16670.
  8. 8. Stathakarou N, Nifakos S, Karlgren K, Konstantinidis ST, Bamidis PD, Pattichis CS, et al. Students’ perceptions on chatbots’ potential and design characteristics in healthcare education. Stud Health Technol Inform. 2020;272:209–12. pmid:32604638
  9. 9. Goodman RS, Patrinely JR, Stone CA Jr, Zimmerman E, Donald RR, Chang SS, et al. Accuracy and reliability of chatbot responses to physician questions. JAMA Netw Open. 2023;6(10):e2336483. pmid:37782499
  10. 10. Temsah O, Khan SA, Chaiah Y, Senjab A, Alhasan K, Jamal A, et al. Overview of early ChatGPT’s presence in medical literature: insights from a hybrid literature review by ChatGPT and human experts. Cureus. 2023;15(4):e37281. pmid:37038381
  11. 11. Chatterjee J, Dethlefs N. This new conversational AI model can be your friend, philosopher, and guide … and even your worst enemy. Patterns (N Y). 2023;4(1):100676. pmid:36699746
  12. 12. Li H, Moon JT, Purkayastha S, Celi LA, Trivedi H, Gichoya JW. Ethics of large language models in medicine and medical research. Lancet Digit Health. 2023;5(6):e333–5. pmid:37120418
  13. 13. Chen M, Zhang B, Cai Z, Seery S, Gonzalez MJ, Ali NM, et al. Acceptance of clinical artificial intelligence among physicians and medical students: a systematic review with cross-sectional survey. Front Med (Lausanne). 2022;9:990604. pmid:36117979
  14. 14. Banerjee M, Chiew D, Patel KT, Johns I, Chappell D, Linton N, et al. The impact of artificial intelligence on clinical education: perceptions of postgraduate trainee doctors in London (UK) and recommendations for trainers. BMC Med Educ. 2021;21(1):429. pmid:34391424
  15. 15. Temsah M-H, Aljamaan F, Malki KH, Alhasan K, Altamimi I, Aljarbou R, et al. ChatGPT and the future of digital health: a study on healthcare workers’ perceptions and expectations. Healthcare (Basel). 2023;11(13):1812. pmid:37444647
  16. 16. Spotnitz M, Idnay B, Gordon ER, Shyu R, Zhang G, Liu C, et al. A survey of clinicians’ views of the utility of large language models. Appl Clin Inform. 2024;15(2):306–12. pmid:38442909
  17. 17. Harris PA, Taylor R, Thielke R, Payne J, Gonzalez N, Conde JG. Research electronic data capture (REDCap)--a metadata-driven methodology and workflow process for providing translational research informatics support. J Biomed Inform. 2009;42(2):377–81. pmid:18929686
  18. 18. Harris PA, Taylor R, Minor BL, Elliott V, Fernandez M, O’Neal L, et al. The REDCap consortium: building an international community of software platform partners. J Biomed Inform. 2019;95:103208. pmid:31078660
  19. 19. Rao SK, Kimball AB, Lehrhoff SR, Hidrue MK, Colton DG, Ferris TG, et al. The impact of administrative burden on academic physicians: results of a hospital-wide physician survey. Acad Med. 2017;92(2):237–43. pmid:28121687
  20. 20. Teubner T, Flath CM, Weinhardt C, van der Aalst W, Hinz O. Welcome to the Era of ChatGPT et al. Bus Inf Syst Eng. 2023;65(2):95–101.
  21. 21. Cascella M, Montomoli J, Bellini V, Bignami E. Evaluating the feasibility of chatgpt in healthcare: an analysis of multiple clinical and research scenarios. J Med Syst. 2023;47(1):33. pmid:36869927
  22. 22. Solmonovich R, Kouba I, Quezada O, Rodriguez-Ayala G, Rojas V, Bonilla K, et al. Artificial Intelligence Generates Proficient Spanish Obstetrics and Gynecology Counseling Templates. AJOG Global Reports. 2024:100400.
  23. 23. Hartman V, Zhang X, Poddar R, Mccarty M, Fortenko A, Sholle E. Developing and evaluating large language model–generated emergency medicine handoff notes. JAMA Network Open. 2024;7(12):e2448723.
  24. 24. Shah NH, Entwistle D, Pfeffer MA. Creation and adoption of large language models in medicine. JAMA. 2023;330(9):866–9. pmid:37548965
  25. 25. Willem T, Krammer S, Böhm A-S, French LE, Hartmann D, Lasser T, et al. Risks and benefits of dermatological machine learning health care applications-an overview and ethical analysis. J Eur Acad Dermatol Venereol. 2022;36(9):1660–8. pmid:35490413
  26. 26. Heinrichs B, Eickhoff SB. Your evidence? Machine learning algorithms for medical diagnosis and prediction. Hum Brain Mapp. 2020;41(6):1435–44. pmid:31804003
  27. 27. Ali S, Akhlaq F, Imran AS, Kastrati Z, Daudpota SM, Moosa M. The enlightening role of explainable artificial intelligence in medical & healthcare domains: a systematic literature review. Comput Biol Med. 2023;166:107555. pmid:37806061
  28. 28. Ribeiro T, Singh S, Guestrin C. Why should I trust you?. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2016;1135–44.
  29. 29. Lundberg SM, Lee S-I. A unified approach to interpreting model predictions. Adv Neural Infor Process Syst. 2017;30.