Figures
Abstract
Background and objective
MyDispense is one of the virtual simulations that has already been established as a suitable alternative for live experiential education in the pharmacy curriculum. However, there are no structured validated questionnaires available to assess the students’ perception while integrating MyDispense with pharmacy practice experiential education. Therefore, the present study aimed to validate a structured questionnaire and use the questionnaire to assess the student perception of various pharmacy practice experiential education.
Methods
Content and construct validity procedure was used to validate the questionnaire. Two hundred students consented to participate in validating the questionnaire. The validated questionnaire assessed the students’ perception of integrating MyDispense with Introductory Pharmacy Practice Experience 2 (IPPE2) and Advanced Pharmacy Practice Experience (APPE) courses. The questionnaire was structured with four domains which were: exercise, instructor, technical, and communication. Each domain carried five items; therefore, the whole questionnaire had 20 items that succeeded in content validity. In the survey, 121 fourth-year and 117 fifth-year Pharm.D. students volunteered to convey their perception of integrating MyDispense with IPPE 2 and APPE, respectively. The survey was conducted before and after the MyDispense exam in both the courses.
Results
The Cronbach’s α and McDonald’s ω coefficients were > 0.8 in all four domains, indicating that the items related to the four domains have good internal consistency. In Exploratory Factor Analysis (EFA), two items were found to cross-load in the exercise domain and removed. Therefore, the EFA proposes 18 items for the confirmatory factor analysis (CFA). In CFA, five fit indices were found to be satisfactory, and this indicates construct was good enough to assess the student perception. In IPPE 2, the pre-test response, the students had significantly higher satisfaction (p < 0.05) with all five items related to the technical domain. In APPE, the students had a significantly (p < 0.05) higher perception of all the items related to the exercise and technical domain in the pre-test compared to the post-test. Therefore, the student’s pre-test feedback allowed the instructor to identify and make the necessary corrections in the exercises to improve the quality exercises.
Citation: Amirthalingam P, Pakkir Mohamed SH, Veeramani VP, Nagoor Thangam MM, Alanazi MF, Dhanasekaran M, et al. (2024) The effectiveness of a structured validated questionnaire to assess student perception with virtual pharmacy simulation in pharmacy practice experiential education. PLoS ONE 19(11): e0314117. https://doi.org/10.1371/journal.pone.0314117
Editor: Kazeem Babatunde Yusuff, Qatar University College of Pharmacy, QATAR
Received: February 28, 2024; Accepted: November 5, 2024; Published: November 21, 2024
Copyright: © 2024 Amirthalingam et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: According to the University policy and local research ethics committee guidelines, the data should not be available in public; however, they may be shared upon request by emailing a_hamdan@ut.edu.sa. The above mentioned email belongs to the authorized person who is responsible for holding all the student-related research data in our institution.
Funding: The authors extend their appreciation to the Deanship of Scientific Research at the University of Tabuk for funding this work through research no. S-1443-0034. The funders had no role in study design, data collection and analysis, decision to publish, or manuscript preparation.
Competing interests: The authors have declared that no competing interests exist.
1. Introduction
Pharmacy practice education usually involves traditional classroom lectures and training in pharmacy; however, technical advancement in recent years has directed the paradigm shift from the conventional classroom to computer-based technology education [1–3]. Pharmacy practice experience education is the best way to prepare the pharmacy student seeking a career as a pharmacist since it offers cognitive, technical, and decision-making skills needed for medication dispensing [4–7]. Pharmacy simulation is a valuable tool in pharmacy practice experience education for training pharmacy students by providing real-world experience [4,5]. However, pharmacy simulation was found to have numerous challenges, including extensive planning, highly expensive simulated pharmacy, and the requirement of many teaching staff for supervision [6]. In this context, virtual simulation has been widely accepted by pharmacy schools since it minimizes the workload of teaching staff, cost, and time while offering pharmacy practice experiences to students [2,5,7].
MyDispense is a digital education platform that enables students to practice the skills of a pharmacist, from novice to highly advanced, in a safe virtual environment that is web-based and highly accessible [7]. The students can learn from their mistakes due to the instant feedback MyDispense in a risk-free environment [8]. Previous studies have established that the MyDispense database helped assess the various student skills regarding dispensing, communication, decision-making, and problem-solving [7–9]. This platform was widely accepted by various pharmacy schools worldwide and implemented in various pharmacy practice experience courses [9]. Although MyDispense plays a vital role in pharmacy practice experience education, recent research addressed numerous barriers regarding implementation, including a lack of training for the staff members in building or updating cases, lack of realism, and challenges in the harmonization of drug nomenclature [10,11].
The implementation of MyDispense is varied among the institutions in terms of exercises and assessment criteria [5,12]. Therefore, assessing students’ perception of MyDispense is essential to ensure effective implementation in every institution. Previous studies already addressed the student perception of integrating MyDispense in various pharmacy courses [7,13–15]. In this context, to the best of our knowledge, structured validated questionnaires are yet to be established [4,7,9]. Therefore, the present study aimed to validate the pioneer multi-dimensional questionnaire for assessing student perception while integrating MyDispense with several pharmacy practice experiential education. Additionally, the study aimed to assess the student perception regarding MyDispense in various pharmacy practice experiential education by using the validated questionnaire.
2. Methods
2.1 Study design and ethical considerations
The study was conducted in two parts. First, it was planned to validate a 5-point multi-dimensional structured Likert-scale questionnaire using content and construct validity procedure. Second, the student perception was assessed with a validated questionnaire in a cross-sectional survey of integrating MyDispense among the various pharmacy practice experiential education. The student volunteers were recruited during the period from 01.03.2022 to 30.12.2023 to integrate MyDispense for the first time at the Faculty of Pharmacy, University of Tabuk, Saudi Arabia. The exercises include patient fact-finding, patient education, answering patient questions, prescription monitoring, labeling, and dispensing (Table 1). Introductory Pharmacy Practice Experience 2 (IPPE 2) and Advanced Pharmacy Practice Experience (APPE) courses were offered in the PharmD curriculum during the fourth and fifth years, respectively.
2.2. Details of exercise and its conductance in MyDispense
The exercises created in MyDispense were based on the course learning outcomes of the courses IPPE2 and APPE. A virtual patient appeared on the computer monitor and sought for his/her prescription to be filled. The exercises in IPPE 2 mainly aimed to test technical skills, including prescription monitoring, labeling, dispensing, and assessing communication skills. Meanwhile, the exercises in APPE also test the technical skills of labeling and dispensing skills; however, the main objective was to test the communication skills in patient fact-finding, patient education, and, importantly, answering patient questions regarding medications. The tutorial and formative assessment case details were attached as S1 File. IPPE2 and APPE courses were conducted for thirteen weeks in the first semester, and MyDispense was integrated for the first four weeks. The students were given a demonstration by the instructor regarding the MyDispense exercise and the rubrics for evaluation in detail in the first week, followed by a tutorial exercise that took place in the second week of the semester. A pre-test survey was conducted immediately after the completion of the tutorial exercise. A formative assessment in the third week, followed by a post-test survey conducted immediately. The actual summative assessment took place in the fourth week and the surveys were not permitted according to the University policy before or after the summative assessment. Students were posted in the pharmacy from the fifth to the thirteenth week for real-world experience. The MyDispense demonstration, tutorial, and formative assessment were carried out in the computer lab with internet facilities at the Faculty of Pharmacy, University of Tabuk.
2.3. Ethical approval and informed consent
The study was approved by the local research ethics committee from the University of Tabuk, Saudi Arabia (Reference number: UT-187-42-2022). Written informed consent was obtained from the student who volunteered to participate in the study. The students were given bonus points for participation in surveys.
2.4 Study participants
Firstly, two hundred students participated in validating the questionnaire. Then, 238 students responded to the validated questionnaire in a cross-sectional survey to address their perception of MyDispense integrated with IPPE2 (n = 121) and APPE (n = 117) courses. In IPPE 2, 50 male (41.32%) and 71 female (58.67%) students participated in the study, and 55 males (47%) and 62 females (53%) were included in the APPE group.
2.5 Components of questionnaire
The questionnaire was structured with four domains, including exercise, instructor, technical, and communication. Each domain has five items; hence, 20 items were included in the 5-point Likert scale questionnaire (Table 2). All four domains have been developed by consensus among the authors of this research based on the IPPE 2 and APPE courses’ learning outcomes, teaching strategies, and instruction methods. Three domains, including exercise, technical, and communication, were developed on the basis of IPPE 2 and APPE courses’ learning outcome domains, including knowledge, cognitive, and skill. The instructor domain consists of teaching strategies and instruction methods for the course specifications of IPPE2 and APPE. The students’ responses were recorded as strongly agree, agree, neutral, disagree, and strongly disagree, which reflects the scores 5,4,3,2 and 1, respectively.
2.6 Validation of questionnaire
2.6.1 Content validity.
Content validity was performed to evaluate how well the questionnaire covered the relevant parts of the construct [16]. The questionnaire was distributed to the five experts in pharmacy practice education outside the institution for content validity. Item-level content validity indexes (I-CVIs) and the averaging of scale-level content validity index (S-CVI/Ave) were included in the content validity. The scores of I-CVIs ≥ 0.78 and S-CVI/Ave ≥ 0.90 were considered excellent content validity [17].
2.6.2 Construct validity.
Cronbach’s α and McDonald’s ω coefficients were used to assess the internal consistency of the questionnaire. Cronbach’s α and McDonald’s ω coefficients of values > 0.9, > 0.7 to ≤ 9, and < 0.7 are considered excellent, good, and poor, respectively [18,19]. The model was constructed with four domains, each with five items initially. Exploratory factor analysis (EFA) was performed using maximum likelihood extraction with a Varimax rotation method [20]. Factor loading > 0.5 was considered an inclusion criterion to retain the items related to their factor, and cross-loaded items with > 0.32 were removed from the construct [21]. Bartlett’s test for sphericity of < 0.05 and Kaiser-Meyer-Olkin–Measuring Sampling Adequacy (KMO-MSA) value ≥ 0.7 were considered acceptable for sampling adequacy [22,23]. The threshold for the cumulative percentage of variance was 50.2% was considered [24]. The robust unweight least square estimation method was used in confirmatory factor analysis (CFA) [25]. According to the recommendations of Hair et al., 2010, more than three fit indices used to establish the model fitness were Chi-square to degrees of freedom ratio (χ2/df), root mean square error of approximation (RMSEA), standardized root mean square residual (SRMR), comparative fit index (CFI), and Tucker Lewis index [26]. χ2/df < 5, RMSEA ≤0.08, SRMR ≤ 0.08, CFI >0.9, and TLI >0.9 were considered a good model fit [27]. The average variance extracted (AVE) ≥ 0.5 and construct reliability for the latent factors (≥ 0.7) were considered for the desired convergent validity [20]. The squared correlation (SC) values of one construct with other constructs are less than the AVE of a specific construct and were considered for the desired discriminant validity [28,29].
2.7 Sample size calculation
For the construct validity procedure, the sample size calculation was based on a 1:5 ratio (number of items: participants) [20]. In the split-half method, the first 100 students were given an odd number of items, and the remaining 100 were given an even number of items in the questionnaire. Therefore, 200 students were recruited to validate the questionnaire. A convenient sampling method was used in the cross-sectional survey.
2.8 Statistical analysis
Statistical Package for the Social Sciences (SPSS version 25.0) was used to perform reliability statistics, factor analysis, and data analysis of the cross-sectional survey. The Mean (Standard Deviation) value of student perception was compared between the pre-test and post-test by using the Wilcoxon signed rank test. P < 0.05 was considered statistically significant.
3. Results
3.1 Validation
The I-CVIs for the relevancy of the questionnaire ranged from 0.8 to 1, and the S-CVI/Ave was >0.9. Hence, I-CVIs and S-CVI/Ave demonstrated that the 20-item questionnaire underlying four factors has excellent content validity. The Cronbach’s α and McDonald’s ω coefficients were > 0.8 in all four factors, indicating that the items related to the four factors have good internal consistency (Table 2). The sample adequacy was confirmed by the KMO-MSA (0.896) and Bartlett’s test for sphericity (p < 0.001). The total cumulative percentage of variance (65.05%) indicates that the proportion of variance explained by the factors was satisfactory. A couple of items with cross-loading values > 0.32 in the exercise were removed from the questionnaire (Table 2). Hence, an 18-item questionnaire was proposed by EFA, and the same was investigated in CFA. The four-factor questionnaire construct was found to have acceptable fit indices, including χ2/df (2.16), RMSEA (0.076), SRMR (0.045), CFI (0.938), and TLI (0.927) (Table 3). The average variance extracted (AVE) of each factor was > 0.5, indicating the convergent validity of the construct (Table 4). The discriminant validity was determined since the AVE of each factor was more than the squared correlations (SC) with the other factors (Table 4). Therefore, an 18-item questionnaire with four factors was constructed successfully to assess the students’ perception regarding virtual simulation.
3.2 Survey
The students were asked to complete the questionnaire before and after the test in a virtual simulation for the courses IPPE 2 and APPE. Overall, the students’ perceptions were good in the pre-and post-tests for both IPPE 2 and APPE courses (Tables 5 and 6). In IPPE 2, the students’ perception of pre-and post-tests had no significant difference regarding exercise, instructor, and communication domains. However, in the pre-test response, the students had significantly higher satisfaction (p < 0.05) with all five items related to the technical domain. In APPE, the students had a significantly higher perception of all the items related to the exercise and technical domain in the pre-test compared to the post-test. Satisfaction was significantly higher in the pre-test regarding three items of the instructor domain about the instructor having adequate experience in MyDispense (p = 0.015), the instructor encouraging me to complete the exercise successfully (p = 0.002), and the instructor being available to clarify my doubts. Three items in the communication domain had significantly higher perceptions in the pre-test, including "I was confident during the interaction with the virtual patient" (p = 0.000), "It seemed like a real-world experience when I engaged with a virtual patient" (p = 0.000), and "The interaction with the virtual patient was useful to complete the session successfully" (p = 0.000).
4. Discussion
The study validated an 18-item structured questionnaire to assess the student perception of virtual pharmacy. It was already well established that MyDispense supports academicians to enhance student learning, increase academic and practical knowledge, and develop essential skills needed to become a pharmacist [9,30,31]. The Pharmacy students’ career option preference as a pharmacist was strongly influenced by their satisfaction with the activities in the Pharmacy curriculum [32]. Therefore, it is important to obtain students’ perceptions periodically to update the activities in the curriculum. Previous studies addressed the student perception with the questionnaire related to the learning, technical, and communication aspects in terms of virtual patients [4,13,15,33]. In addition, our questionnaire addresses the students’ perceptions about the suitability of exercise for the course and clarity of exercise and assessment criteria. Also, the questionnaire can know the student’s perception regarding their instructor in various aspects, including knowledge of the exercise, experience in MyDispense, demonstration skills, motivation of students, and availability to sort out any difficulties. Successful implementation of the virtual simulation exercise depends on the resources available in the institution [34]. Integrating virtual pharmacy simulation among experiential education needs a careful selection of exercises and strategic implementation of them within the courses [30]. Student perception of instructors is strongly related to their satisfaction with online learning, and motivation of students toward virtual platforms will help to achieve the learning outcomes [35–37]. Hopefully, the additional domains in the questionnaire will help the academicians improve the quality of creating and implementing the exercises and instructional approach through the feedback from the students.
The questionnaire was implemented in pre-test and post-test in both IPPE 2 and APPE. Overall, the student’s perception was found to be good regarding integrating MyDispense with IPPE 2 and APPE in terms of all the domains in the questionnaire. In IPPE2, the students found difficulty regarding technical aspects during the test, reflected in their post-test responses. To the best of our knowledge, there are no reports of psychological or physiological factors, if any, deteriorating student performance while appearing with MyDispense for the first time. However, previous researchers reported that virtual simulation reduces anxiety among occupational therapy students [38]; however, another author published a conflicting report that virtual simulation was capable of causing psychological and physiological stress among medical and nursing students [39]. This study warrants future studies to investigate the correlation between the level of stress and anxiety with performance while integrating MyDispense in introductory experiential education. The exercises in the tutorial and formative assessment might have differences in complexity level (S1 File). In this regard, the complexity of the exercise could be the potential confounding factor for the significantly lower perception of the students in the post-test survey. Therefore, the academicians need to pay more attention to maintain the complexity level of exercise in both tutorial and evaluation.
The usefulness of MyDispense in improving communication skills has already been established regarding comprehensive patient-fact findings, and it also facilitates the students to prepare the counseling points in various practice set-ups [9]. The exercises in APPE mainly focussed on testing the skills related to the communication aspects followed by technical skills. The student perception related to all the items in the exercise and three items related to the instructor domain were significantly lower in the post-test survey. This feedback from the students will allow the academicians to improve the quality of exercises and the involvement of instructors [30,35–37]. The students highly perceived the MyDispense pre-test in all the items of the technical domain and three items in communication aspects. It might be due to the psychological or physiological stress associated with the students during the exam [39]. The post-test student responses imply that the student needs more motivation to answer patient questions and dispense multiple medications in APPE at a given time. Henceforth, the academicians can able to improve the quality of exercises and improve their participation to enhance the student’s face-to-face real-world experience in terms of handling complex situations in pharmacy practice [32,40].
4.1 Strength and limitations
The study has come up with a new 18-item validated questionnaire for the first time, and the questionnaire can be implemented in both introductory and advanced experiential education. However, the study has the following limitations: 1. A limited sample size may be unable to judge all the aspects of exercises, instructor, technical, and communication, 2. The study was conducted in one site, and the results cannot be generalized. 3. The questionnaire was validated by acquiring students from a single institution; it could have been obtained from students of different Universities, and 4. Both pre-test and post-test surveys took place after only one exercise might not be enough to reflect student satisfaction accurately.
5. Conclusion
This study supplies a validated 18-item questionnaire to assess the student perception of the integration of MyDispense with introductory and advanced experiential education. This questionnaire can help the academicians to enhance the quality of exercises and instructors by acquiring the student’s perceptions. The complexity of exercises reflects the student satisfaction; hence, the instructors need to be well-prepared with good-quality exercises and also need to encourage the students to perform well while integrating MyDispense with advanced-level experiential education.
Acknowledgments
We are thankful to Monash University, Australia, for partnering with the University of Tabuk to utilize MyDispense to train pharmacy students in various aspects (https://info.mydispense.monash.edu/community/partner-institutions/).
References
- 1. Gharib AM, Bindoff IK, Peterson GM, Salahudeen MS. Computer-Based Simulators in Pharmacy Practice Education: A Systematic Narrative Review. Pharmacy (Basel). 2023;11(1):8. Published 2023 Jan 2. pmid:36649018
- 2. Lim AS, Lee SWH. Is Technology Enhanced Learning Cost-effective to Improve Skills?: The Monash Objective Structured Clinical Examination Virtual Experience. Simul Healthc. 2022;17(2):131–135. pmid:33273417
- 3. Hamilton LA, Suda KJ, Heidel RE, McDonough SLK, Hunt ME, Franks AS. The role of online learning in pharmacy education: A nationwide survey of student pharmacists. Curr Pharm Teach Learn. 2020;12(6):614–625. pmid:32482262
- 4. Korayem GB, Alboghdadly AM. Integrating simulation into advanced pharmacy practice experience curriculum: An innovative approach to training. Saudi Pharm J. 2020;28(7):837–843. pmid:32647485
- 5. Lin K, Travlos DV, Wadelin JW, Vlasses PH. Simulation and introductory pharmacy practice experiences. Am J Pharm Educ. 2011;75(10):209. pmid:22345728
- 6. Katoue MG, Ker J. Simulation for Continuing Pharmacy Education: Development and Implementation of a Simulation-Based Workshop on Medicines Reconciliation for Pharmacists. J Contin Educ Health Prof. 2019;39(3):185–193. pmid:31166221
- 7. McDowell J, Styles K, Sewell K, et al. A Simulated Learning Environment for Teaching Medicine Dispensing Skills. Am J Pharm Educ. 2016;80(1):11. pmid:26941437
- 8. Mak V, Fitzgerald J, Holle L, Vordenberg SE, Kebodeaux C. Meeting pharmacy educational outcomes through effective use of the virtual simulation MyDispense. Curr Pharm Teach Learn. 2021;13(7):739–742. pmid:34074500
- 9. Khera HK, Mannix E, Moussa R, Mak V. MyDispense simulation in pharmacy education: a scoping review. J Pharm Policy Pract. 2023;16(1):110. pmid:37770985
- 10. Seybert AL, Smithburger PL, Benedict NJ, Kobulinsky LR, Kane-Gill SL, Coons JC. Evidence for simulation in pharmacy education. J Am Coll Clin Pharm. 2019;2(6):686–692.
- 11. Yuan HB, Williams BA, Fang JB, Ye QH. A systematic review of selected evidence on improving knowledge and skills through high-fidelity simulation. Nurse Educ Today. 2012;32(3):294–298. pmid:21862186
- 12. Presado MHCV, Colaço S, Rafael H, et al. Learning with High Fidelity Simulation. Aprender com a Simulação de Alta Fidelidade. Cien Saude Colet. 2018;23(1):51–59. pmid:29267811
- 13. Amirthalingam P, Hamdan AM, Veeramani V.P, A Sayed Ali M. A comparison between student performances on objective structured clinical examination and virtual simulation. Pharm. Educ. 2022;22(1):466–473.
- 14. Lucas C, Williams K, Bajorek B. Virtual Pharmacy Programs to Prepare Pharmacy Students for Community and Hospital Placements. Am J Pharm Educ. 2019;83(10):7011. pmid:32001870
- 15. Shin J, Tabatabai D, Boscardin C, Ferrone M, Brock T. Integration of a Community Pharmacy Simulation Program into a Therapeutics Course. Am J Pharm Educ. 2018;82(1):6189. pmid:29491500
- 16. Zamanzadeh V, Ghahramanian A, Rassouli M, Abbaszadeh A, Alavi-Majd H, Nikanfar AR. Design and Implementation Content Validity Study: Development of an instrument for measuring Patient-Centered Communication. J Caring Sci. 2015;4(2):165–178. Published 2015 Jun 1. pmid:26161370
- 17. Polit DF, Beck CT, Owen SV. Is the CVI an acceptable indicator of content validity? Appraisal and recommendations. Res Nurs Health. 2007;30(4):459–467. pmid:17654487
- 18. Deng L, Chan W. Testing the Difference Between Reliability Coefficients Alpha and Omega. Educ Psychol Meas. 2017;77(2):185–203. pmid:29795909
- 19. Viladrich C, Angulo-Brunet A, Doval E. A Journey around alpha and omega to estimate internal consistency reliability. Anales de Psicología, 2017;33(3):755–782.
- 20. Mya KS, Zaw KK, Mya KM. Developing and validating a questionnaire to assess an individual’s perceived risk of four major non-communicable diseases in Myanmar. PLoS One. 2021;16(4):e0234281. Published 2021 Apr 27. pmid:33905409
- 21. Tavakol M, Wetzel A. Factor Analysis: a means for theory and instrument development in support of construct validity. Int J Med Educ. 2020;11:245–247. Published 2020 Nov 6. pmid:33170146
- 22. Bartlett MS. A Note on the Multiplying Factors for Various χ2 Approximations. J R Stat Soc Series B. 1954;16:296–8.
- 23. Kaiser HF. An index of factorial simplicity. Psychometrika. 1974;39,31–36.
- 24. Peterson RA. A Meta-Analysis of Variance Accounted for and Factor Loadings in Exploratory Factor Analysis. Marketing Letters. 2000:11:261–275.
- 25. Kogar H, Yılmaz Kogar E. Comparison of Different Estimation Methods for Categorical and Ordinal Data in Confirmatory Factor Analysis. J. Meas. Eval. 2016;6(2): 0–0.
- 26.
Hair JF, Black WC, Babin BJ, Anderson RE. Multivariate Data Analysis: A Global Perspective. 7th Edition, 2010; Pearson Education: Upper Saddle Rive.
- 27. Hou D, Al-Tabbaa A, Chen H, Mamic I. Factor analysis and structural equation modelling of sustainable behaviour in contaminated land remediation. J Clean Prod. 2014;84:439–49. https://select-statistics.co.uk/calculators/sample-size-calculator-two-means/.
- 28. Costello AB, Osborne JW. Best practices in exploratory factor analysis: Four recommendations for getting the most from your analysis. Pract. Assess. Res. Evaluation. 2005;10:1–9.
- 29. Fabrigar LR, Wegener DT, MacCallum RC, Strahan EJ. Evaluating the use of exploratory factor analysis in psychological research. Psychol. Methods. 1999;4(3):272–299.
- 30. Phanudulkitti C, Puengrung S, Meepong R, Vanderboll K, Farris KB, Vordenberg SE. A systematic review on the use of virtual patient and computer-based simulation for experiential pharmacy education. Explor Res Clin Soc Pharm. 2023;11:100316. Published 2023 Aug 7. pmid:37635840
- 31. Al-Diery T, Hejazi T, Al-Qahtani N, ElHajj M, Rachid O, Jaam M. Evaluating the use of virtual simulation training to support pharmacy students’ competency development in conducting dispensing tasks. Curr Pharm Teach Learn. Published online September 5, 2024. pmid:39241581
- 32. Alnahar SA, Mamiya KT, John C, Bader L, Bates I. Experience with pharmacy academic programmes and career aspirations of pharmacy students and young pharmacists-an international cross-sectional study. BMC Med Educ. 2022;22(1):444. Published 2022 Jun 8. pmid:35676676
- 33. Johnson AE, Barrack J, Fitzgerald JM, Sobieraj DM, Holle LM. Integration of a Virtual Dispensing Simulator "MyDispense" in an Experiential Education Program to Prepare Students for Community Introductory Pharmacy Practice Experience. Pharmacy (Basel). 2021;9(1):48. Published 2021 Feb 27. pmid:33673541
- 34. Wald DA. Cripe J, Garcia P. Incorporating a virtual simulation exercise into the preclerkship undergraduate curriculum. Acad Emerg Med. 2021;28(SUPPL 1):S399–S400, 2021.
- 35. Mills JM, VanAtta CN, Hendershot RS, Rao S. Pharmacy Students’ Perceptions of Remote versus Face-to-Face Learning Experience. Pharmacy (Basel). 2023;11(3):97. Published 2023 Jun 8. pmid:37368423
- 36. Wang R, Han J, Liu C, Xu H. How Do University Students’ Perceptions of the Instructor’s Role Influence Their Learning Outcomes and Satisfaction in Cloud-Based Virtual Classrooms During the COVID-19 Pandemic?. Front Psychol. 2021;12:627443. Published 2021 Apr 16. pmid:33935878
- 37. Lee SJ. Srinivasan S. Trail T. Lewis D. Lopez S. Examining the relationship among student perception of support, course satisfaction, and learning outcomes in online learning. Internet High. Educ. 2011;14:158–163.
- 38. Concannon BJ, Esmail S, Roduta Roberts M. Immersive Virtual Reality for the Reduction of State Anxiety in Clinical Interview Exams: Prospective Cohort Study. JMIR Serious Games. 2020;8(3):e18313. Published 2020 Jul 9. pmid:32673223
- 39. Liaw SY, Sutini , Chua WL, et al. Desktop Virtual Reality Versus Face-to-Face Simulation for Team-Training on Stress Levels and Performance in Clinical Deterioration: a Randomised Controlled Trial. J Gen Intern Med. 2023;38(1):67–73. pmid:35501626
- 40. Ambroziak K., Ibrahim N., Marshall V. D., & Kelling S. E. (2018). Virtual simulation to personalize student learning in a required pharmacy course. Curr Pharm Teach Learn. 10(6), 750–756. pmid:30025776